2024-12-09 20:52:59,104 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-09 20:52:59,114 main DEBUG Took 0.008939 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-09 20:52:59,114 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-09 20:52:59,115 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-09 20:52:59,116 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-09 20:52:59,116 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,131 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-09 20:52:59,142 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,143 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,144 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,145 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,145 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,146 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,147 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,147 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,148 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,148 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,149 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,150 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,151 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,151 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,151 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,152 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,152 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,153 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,153 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,154 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,154 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,155 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,155 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,155 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 20:52:59,156 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,156 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-09 20:52:59,158 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 20:52:59,160 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-09 20:52:59,162 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-09 20:52:59,163 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-09 20:52:59,164 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-09 20:52:59,165 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-09 20:52:59,176 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-09 20:52:59,179 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-09 20:52:59,182 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-09 20:52:59,182 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-09 20:52:59,183 main DEBUG createAppenders(={Console}) 2024-12-09 20:52:59,184 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-09 20:52:59,184 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-09 20:52:59,184 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-09 20:52:59,185 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-09 20:52:59,185 main DEBUG OutputStream closed 2024-12-09 20:52:59,186 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-09 20:52:59,186 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-09 20:52:59,186 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-09 20:52:59,258 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-09 20:52:59,260 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-09 20:52:59,261 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-09 20:52:59,261 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-09 20:52:59,262 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-09 20:52:59,262 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-09 20:52:59,263 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-09 20:52:59,263 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-09 20:52:59,263 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-09 20:52:59,263 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-09 20:52:59,263 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-09 20:52:59,264 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-09 20:52:59,264 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-09 20:52:59,264 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-09 20:52:59,265 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-09 20:52:59,265 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-09 20:52:59,265 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-09 20:52:59,266 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-09 20:52:59,268 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-09 20:52:59,268 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-09 20:52:59,268 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-09 20:52:59,269 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-09T20:52:59,282 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-09 20:52:59,284 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-09 20:52:59,284 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-09T20:52:59,496 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33 2024-12-09T20:52:59,517 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416, deleteOnExit=true 2024-12-09T20:52:59,518 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/test.cache.data in system properties and HBase conf 2024-12-09T20:52:59,519 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.tmp.dir in system properties and HBase conf 2024-12-09T20:52:59,519 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir in system properties and HBase conf 2024-12-09T20:52:59,520 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-09T20:52:59,520 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-09T20:52:59,520 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-09T20:52:59,597 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-09T20:52:59,679 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-09T20:52:59,683 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-09T20:52:59,683 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-09T20:52:59,683 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-09T20:52:59,684 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-09T20:52:59,684 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-09T20:52:59,685 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-09T20:52:59,685 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-09T20:52:59,685 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-09T20:52:59,686 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-09T20:52:59,686 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/nfs.dump.dir in system properties and HBase conf 2024-12-09T20:52:59,686 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/java.io.tmpdir in system properties and HBase conf 2024-12-09T20:52:59,686 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-09T20:52:59,687 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-09T20:52:59,687 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-09T20:53:00,731 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-09T20:53:00,791 INFO [Time-limited test {}] log.Log(170): Logging initialized @2261ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-09T20:53:00,851 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:00,906 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T20:53:00,928 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T20:53:00,928 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T20:53:00,929 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-09T20:53:00,940 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:00,943 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@760c69c0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir/,AVAILABLE} 2024-12-09T20:53:00,945 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@ce709a8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T20:53:01,119 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@62d6efd9{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/java.io.tmpdir/jetty-localhost-35581-hadoop-hdfs-3_4_1-tests_jar-_-any-12801569820756107838/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-09T20:53:01,126 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@353d35a1{HTTP/1.1, (http/1.1)}{localhost:35581} 2024-12-09T20:53:01,127 INFO [Time-limited test {}] server.Server(415): Started @2597ms 2024-12-09T20:53:01,658 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:01,664 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T20:53:01,665 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T20:53:01,665 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T20:53:01,665 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-09T20:53:01,666 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3a5de9e4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir/,AVAILABLE} 2024-12-09T20:53:01,666 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@69893329{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T20:53:01,757 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1b97a472{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/java.io.tmpdir/jetty-localhost-38837-hadoop-hdfs-3_4_1-tests_jar-_-any-7412001360065441852/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:01,758 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3722a29b{HTTP/1.1, (http/1.1)}{localhost:38837} 2024-12-09T20:53:01,758 INFO [Time-limited test {}] server.Server(415): Started @3229ms 2024-12-09T20:53:01,801 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T20:53:01,893 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:01,897 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T20:53:01,898 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T20:53:01,898 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T20:53:01,899 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-09T20:53:01,899 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@510fec09{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir/,AVAILABLE} 2024-12-09T20:53:01,900 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@40eb7053{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T20:53:02,009 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@353955e9{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/java.io.tmpdir/jetty-localhost-41731-hadoop-hdfs-3_4_1-tests_jar-_-any-16022759096976957202/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:02,010 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11738cd8{HTTP/1.1, (http/1.1)}{localhost:41731} 2024-12-09T20:53:02,010 INFO [Time-limited test {}] server.Server(415): Started @3481ms 2024-12-09T20:53:02,012 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T20:53:02,042 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:02,047 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T20:53:02,049 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T20:53:02,050 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T20:53:02,050 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-09T20:53:02,051 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@16cd567f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir/,AVAILABLE} 2024-12-09T20:53:02,051 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5822645a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T20:53:02,144 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3114ae69{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/java.io.tmpdir/jetty-localhost-39099-hadoop-hdfs-3_4_1-tests_jar-_-any-7885705307311321569/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:02,145 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3c70a874{HTTP/1.1, (http/1.1)}{localhost:39099} 2024-12-09T20:53:02,145 INFO [Time-limited test {}] server.Server(415): Started @3615ms 2024-12-09T20:53:02,147 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T20:53:03,454 WARN [Thread-124 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data4/current/BP-1840173676-172.17.0.2-1733777580185/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:03,454 WARN [Thread-123 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data1/current/BP-1840173676-172.17.0.2-1733777580185/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:03,454 WARN [Thread-125 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data2/current/BP-1840173676-172.17.0.2-1733777580185/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:03,454 WARN [Thread-122 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data3/current/BP-1840173676-172.17.0.2-1733777580185/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:03,494 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T20:53:03,494 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T20:53:03,538 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xa511eff469f70b22 with lease ID 0x72db93d7da36ff6d: Processing first storage report for DS-c2e178b9-c94c-4467-b7d3-5489432fe387 from datanode DatanodeRegistration(127.0.0.1:36667, datanodeUuid=9841445a-8140-4932-8cc1-7dc15f099dc2, infoPort=43477, infoSecurePort=0, ipcPort=41459, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185) 2024-12-09T20:53:03,540 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa511eff469f70b22 with lease ID 0x72db93d7da36ff6d: from storage DS-c2e178b9-c94c-4467-b7d3-5489432fe387 node DatanodeRegistration(127.0.0.1:36667, datanodeUuid=9841445a-8140-4932-8cc1-7dc15f099dc2, infoPort=43477, infoSecurePort=0, ipcPort=41459, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2024-12-09T20:53:03,540 WARN [Thread-143 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data6/current/BP-1840173676-172.17.0.2-1733777580185/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:03,540 WARN [Thread-142 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data5/current/BP-1840173676-172.17.0.2-1733777580185/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:03,540 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x842054ea05b3fdc9 with lease ID 0x72db93d7da36ff6e: Processing first storage report for DS-e22c6745-d9a8-48a2-b4b1-814705d31b2e from datanode DatanodeRegistration(127.0.0.1:38901, datanodeUuid=0af9d13e-06ac-4a26-9f56-90d344b0197a, infoPort=33261, infoSecurePort=0, ipcPort=37219, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185) 2024-12-09T20:53:03,540 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x842054ea05b3fdc9 with lease ID 0x72db93d7da36ff6e: from storage DS-e22c6745-d9a8-48a2-b4b1-814705d31b2e node DatanodeRegistration(127.0.0.1:38901, datanodeUuid=0af9d13e-06ac-4a26-9f56-90d344b0197a, infoPort=33261, infoSecurePort=0, ipcPort=37219, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:03,541 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xa511eff469f70b22 with lease ID 0x72db93d7da36ff6d: Processing first storage report for DS-1c334978-00ba-49c7-a1d3-72c18b56baa7 from datanode DatanodeRegistration(127.0.0.1:36667, datanodeUuid=9841445a-8140-4932-8cc1-7dc15f099dc2, infoPort=43477, infoSecurePort=0, ipcPort=41459, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185) 2024-12-09T20:53:03,541 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa511eff469f70b22 with lease ID 0x72db93d7da36ff6d: from storage DS-1c334978-00ba-49c7-a1d3-72c18b56baa7 node DatanodeRegistration(127.0.0.1:36667, datanodeUuid=9841445a-8140-4932-8cc1-7dc15f099dc2, infoPort=43477, infoSecurePort=0, ipcPort=41459, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:03,541 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x842054ea05b3fdc9 with lease ID 0x72db93d7da36ff6e: Processing first storage report for DS-be2aaa9b-6c0f-46bf-8f0a-c50e843d4f8a from datanode DatanodeRegistration(127.0.0.1:38901, datanodeUuid=0af9d13e-06ac-4a26-9f56-90d344b0197a, infoPort=33261, infoSecurePort=0, ipcPort=37219, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185) 2024-12-09T20:53:03,541 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x842054ea05b3fdc9 with lease ID 0x72db93d7da36ff6e: from storage DS-be2aaa9b-6c0f-46bf-8f0a-c50e843d4f8a node DatanodeRegistration(127.0.0.1:38901, datanodeUuid=0af9d13e-06ac-4a26-9f56-90d344b0197a, infoPort=33261, infoSecurePort=0, ipcPort=37219, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:03,557 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T20:53:03,563 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x33f5ed042d14c0d2 with lease ID 0x72db93d7da36ff6f: Processing first storage report for DS-f5161ea3-c80e-42e5-8de0-a41d9afc7c11 from datanode DatanodeRegistration(127.0.0.1:34835, datanodeUuid=40225027-bbd4-49de-98d7-4b182bc0f27b, infoPort=34993, infoSecurePort=0, ipcPort=45951, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185) 2024-12-09T20:53:03,563 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x33f5ed042d14c0d2 with lease ID 0x72db93d7da36ff6f: from storage DS-f5161ea3-c80e-42e5-8de0-a41d9afc7c11 node DatanodeRegistration(127.0.0.1:34835, datanodeUuid=40225027-bbd4-49de-98d7-4b182bc0f27b, infoPort=34993, infoSecurePort=0, ipcPort=45951, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:03,563 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x33f5ed042d14c0d2 with lease ID 0x72db93d7da36ff6f: Processing first storage report for DS-871f9fd9-0fb3-4893-8c76-c450ecdabf58 from datanode DatanodeRegistration(127.0.0.1:34835, datanodeUuid=40225027-bbd4-49de-98d7-4b182bc0f27b, infoPort=34993, infoSecurePort=0, ipcPort=45951, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185) 2024-12-09T20:53:03,563 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x33f5ed042d14c0d2 with lease ID 0x72db93d7da36ff6f: from storage DS-871f9fd9-0fb3-4893-8c76-c450ecdabf58 node DatanodeRegistration(127.0.0.1:34835, datanodeUuid=40225027-bbd4-49de-98d7-4b182bc0f27b, infoPort=34993, infoSecurePort=0, ipcPort=45951, storageInfo=lv=-57;cid=testClusterID;nsid=1567328600;c=1733777580185), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:03,657 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33 2024-12-09T20:53:03,726 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-09T20:53:03,773 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=161, OpenFileDescriptor=393, MaxFileDescriptor=1048576, SystemLoadAverage=239, ProcessCount=11, AvailableMemoryMB=4192 2024-12-09T20:53:03,775 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-09T20:53:03,781 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-09T20:53:03,863 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/zookeeper_0, clientPort=53435, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-09T20:53:03,871 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=53435 2024-12-09T20:53:03,880 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:03,882 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:03,958 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:03,958 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:04,001 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:46348 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:34835:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46348 dst: /127.0.0.1:34835 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:04,021 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-09T20:53:04,421 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:04,434 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147 with version=8 2024-12-09T20:53:04,434 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/hbase-staging 2024-12-09T20:53:04,507 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-09T20:53:04,738 INFO [Time-limited test {}] client.ConnectionUtils(128): master/00c23a7251f8:0 server-side Connection retries=45 2024-12-09T20:53:04,746 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:04,746 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:04,750 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T20:53:04,750 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:04,751 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T20:53:04,861 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-09T20:53:04,908 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-09T20:53:04,915 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-09T20:53:04,918 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T20:53:04,939 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 84911 (auto-detected) 2024-12-09T20:53:04,940 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-09T20:53:04,955 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36171 2024-12-09T20:53:04,973 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:36171 connecting to ZooKeeper ensemble=127.0.0.1:53435 2024-12-09T20:53:05,093 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:361710x0, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T20:53:05,095 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:36171-0x1000c5a75aa0000 connected 2024-12-09T20:53:05,190 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,193 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,203 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:05,208 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147, hbase.cluster.distributed=false 2024-12-09T20:53:05,228 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T20:53:05,232 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36171 2024-12-09T20:53:05,233 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36171 2024-12-09T20:53:05,233 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36171 2024-12-09T20:53:05,236 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36171 2024-12-09T20:53:05,236 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36171 2024-12-09T20:53:05,321 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c23a7251f8:0 server-side Connection retries=45 2024-12-09T20:53:05,322 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,322 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,322 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T20:53:05,322 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,323 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T20:53:05,325 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T20:53:05,326 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T20:53:05,327 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39315 2024-12-09T20:53:05,328 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39315 connecting to ZooKeeper ensemble=127.0.0.1:53435 2024-12-09T20:53:05,329 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,332 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:393150x0, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T20:53:05,347 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:05,347 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39315-0x1000c5a75aa0001 connected 2024-12-09T20:53:05,350 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T20:53:05,356 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T20:53:05,359 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T20:53:05,365 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T20:53:05,366 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39315 2024-12-09T20:53:05,366 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39315 2024-12-09T20:53:05,366 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39315 2024-12-09T20:53:05,367 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39315 2024-12-09T20:53:05,367 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39315 2024-12-09T20:53:05,381 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c23a7251f8:0 server-side Connection retries=45 2024-12-09T20:53:05,381 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,381 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,382 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T20:53:05,382 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,382 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T20:53:05,382 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T20:53:05,383 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T20:53:05,384 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:45461 2024-12-09T20:53:05,385 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:45461 connecting to ZooKeeper ensemble=127.0.0.1:53435 2024-12-09T20:53:05,386 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,389 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:454610x0, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T20:53:05,399 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:454610x0, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:05,399 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45461-0x1000c5a75aa0002 connected 2024-12-09T20:53:05,400 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T20:53:05,401 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T20:53:05,402 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T20:53:05,404 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T20:53:05,405 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45461 2024-12-09T20:53:05,407 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45461 2024-12-09T20:53:05,407 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45461 2024-12-09T20:53:05,411 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45461 2024-12-09T20:53:05,412 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45461 2024-12-09T20:53:05,426 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c23a7251f8:0 server-side Connection retries=45 2024-12-09T20:53:05,426 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,426 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,426 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T20:53:05,426 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:05,427 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T20:53:05,427 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T20:53:05,427 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T20:53:05,428 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:40865 2024-12-09T20:53:05,429 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:40865 connecting to ZooKeeper ensemble=127.0.0.1:53435 2024-12-09T20:53:05,430 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,432 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:408650x0, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T20:53:05,441 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:408650x0, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:05,442 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:40865-0x1000c5a75aa0003 connected 2024-12-09T20:53:05,442 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T20:53:05,443 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T20:53:05,443 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T20:53:05,445 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T20:53:05,448 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40865 2024-12-09T20:53:05,448 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40865 2024-12-09T20:53:05,449 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40865 2024-12-09T20:53:05,450 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40865 2024-12-09T20:53:05,450 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40865 2024-12-09T20:53:05,464 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;00c23a7251f8:36171 2024-12-09T20:53:05,465 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/00c23a7251f8,36171,1733777584597 2024-12-09T20:53:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:05,474 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/00c23a7251f8,36171,1733777584597 2024-12-09T20:53:05,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T20:53:05,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:05,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T20:53:05,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T20:53:05,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:05,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:05,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:05,505 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-09T20:53:05,506 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/00c23a7251f8,36171,1733777584597 from backup master directory 2024-12-09T20:53:05,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:05,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:05,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/00c23a7251f8,36171,1733777584597 2024-12-09T20:53:05,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:05,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:05,516 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T20:53:05,516 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=00c23a7251f8,36171,1733777584597 2024-12-09T20:53:05,518 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-09T20:53:05,520 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-09T20:53:05,578 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/hbase.id] with ID: 9a9faef4-5ccb-4487-a0e3-789175cc03b6 2024-12-09T20:53:05,578 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/.tmp/hbase.id 2024-12-09T20:53:05,585 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:05,585 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:05,588 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:46378 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:34835:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46378 dst: /127.0.0.1:34835 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:05,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-09T20:53:05,595 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:05,595 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/.tmp/hbase.id]:[hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/hbase.id] 2024-12-09T20:53:05,640 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:05,644 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-09T20:53:05,664 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 18ms. 2024-12-09T20:53:05,672 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:05,672 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:05,672 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:05,672 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:05,684 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:05,684 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:05,687 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:35120 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:38901:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35120 dst: /127.0.0.1:38901 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:05,692 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-09T20:53:05,694 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:05,708 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-09T20:53:05,710 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-09T20:53:05,714 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T20:53:05,740 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:05,740 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:05,742 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:46404 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:34835:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46404 dst: /127.0.0.1:34835 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:05,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-09T20:53:05,748 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:05,763 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store 2024-12-09T20:53:05,776 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:05,777 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:05,780 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:32928 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:36667:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:32928 dst: /127.0.0.1:36667 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:05,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-09T20:53:05,785 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:05,789 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-09T20:53:05,792 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:05,793 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-09T20:53:05,793 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:05,794 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:05,795 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-09T20:53:05,795 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:05,795 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:05,796 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733777585793Disabling compacts and flushes for region at 1733777585793Disabling writes for close at 1733777585795 (+2 ms)Writing region close event to WAL at 1733777585795Closed at 1733777585795 2024-12-09T20:53:05,798 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/.initializing 2024-12-09T20:53:05,798 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/WALs/00c23a7251f8,36171,1733777584597 2024-12-09T20:53:05,806 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-09T20:53:05,820 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C36171%2C1733777584597, suffix=, logDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/WALs/00c23a7251f8,36171,1733777584597, archiveDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/oldWALs, maxLogs=10 2024-12-09T20:53:05,854 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/WALs/00c23a7251f8,36171,1733777584597/00c23a7251f8%2C36171%2C1733777584597.1733777585825, exclude list is [], retry=0 2024-12-09T20:53:05,871 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:05,872 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:36667,DS-c2e178b9-c94c-4467-b7d3-5489432fe387,DISK] 2024-12-09T20:53:05,872 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34835,DS-f5161ea3-c80e-42e5-8de0-a41d9afc7c11,DISK] 2024-12-09T20:53:05,872 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38901,DS-e22c6745-d9a8-48a2-b4b1-814705d31b2e,DISK] 2024-12-09T20:53:05,875 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-09T20:53:05,915 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/WALs/00c23a7251f8,36171,1733777584597/00c23a7251f8%2C36171%2C1733777584597.1733777585825 2024-12-09T20:53:05,916 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:33261:33261),(127.0.0.1/127.0.0.1:34993:34993),(127.0.0.1/127.0.0.1:43477:43477)] 2024-12-09T20:53:05,916 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-09T20:53:05,917 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:05,919 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,920 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,951 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,971 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-09T20:53:05,974 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:05,976 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:05,976 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,979 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-09T20:53:05,980 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:05,981 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T20:53:05,981 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,983 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-09T20:53:05,983 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:05,984 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T20:53:05,984 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,986 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-09T20:53:05,987 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:05,987 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T20:53:05,988 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,991 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,992 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,996 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,997 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:05,999 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T20:53:06,003 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:06,009 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T20:53:06,010 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74089012, jitterRate=0.10401231050491333}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T20:53:06,015 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733777585931Initializing all the Stores at 1733777585933 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777585933Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777585934 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777585934Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777585934Cleaning up temporary data from old regions at 1733777585997 (+63 ms)Region opened successfully at 1733777586015 (+18 ms) 2024-12-09T20:53:06,017 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-09T20:53:06,047 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@276d01f5, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c23a7251f8/172.17.0.2:0 2024-12-09T20:53:06,072 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-09T20:53:06,081 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-09T20:53:06,081 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-09T20:53:06,083 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-09T20:53:06,085 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-09T20:53:06,089 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 4 msec 2024-12-09T20:53:06,089 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-09T20:53:06,110 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-09T20:53:06,117 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-09T20:53:06,167 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-09T20:53:06,171 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-09T20:53:06,174 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-09T20:53:06,187 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-09T20:53:06,189 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-09T20:53:06,193 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-09T20:53:06,198 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-09T20:53:06,200 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-09T20:53:06,209 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-09T20:53:06,233 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-09T20:53:06,240 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-09T20:53:06,251 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:06,251 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:06,251 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:06,251 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:06,251 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,251 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,251 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,251 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,254 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=00c23a7251f8,36171,1733777584597, sessionid=0x1000c5a75aa0000, setting cluster-up flag (Was=false) 2024-12-09T20:53:06,283 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,283 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,283 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,283 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,315 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-09T20:53:06,320 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=00c23a7251f8,36171,1733777584597 2024-12-09T20:53:06,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:06,378 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-09T20:53:06,382 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=00c23a7251f8,36171,1733777584597 2024-12-09T20:53:06,390 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-09T20:53:06,451 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-09T20:53:06,456 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(746): ClusterId : 9a9faef4-5ccb-4487-a0e3-789175cc03b6 2024-12-09T20:53:06,456 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(746): ClusterId : 9a9faef4-5ccb-4487-a0e3-789175cc03b6 2024-12-09T20:53:06,456 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(746): ClusterId : 9a9faef4-5ccb-4487-a0e3-789175cc03b6 2024-12-09T20:53:06,458 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T20:53:06,458 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T20:53:06,458 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T20:53:06,460 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-09T20:53:06,466 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-09T20:53:06,471 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 00c23a7251f8,36171,1733777584597 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-09T20:53:06,485 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T20:53:06,485 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T20:53:06,485 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T20:53:06,485 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T20:53:06,485 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T20:53:06,485 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T20:53:06,486 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/00c23a7251f8:0, corePoolSize=5, maxPoolSize=5 2024-12-09T20:53:06,486 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/00c23a7251f8:0, corePoolSize=5, maxPoolSize=5 2024-12-09T20:53:06,486 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/00c23a7251f8:0, corePoolSize=5, maxPoolSize=5 2024-12-09T20:53:06,486 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/00c23a7251f8:0, corePoolSize=5, maxPoolSize=5 2024-12-09T20:53:06,486 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/00c23a7251f8:0, corePoolSize=10, maxPoolSize=10 2024-12-09T20:53:06,486 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,486 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/00c23a7251f8:0, corePoolSize=2, maxPoolSize=2 2024-12-09T20:53:06,487 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,488 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733777616488 2024-12-09T20:53:06,490 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-09T20:53:06,491 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-09T20:53:06,492 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T20:53:06,493 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-09T20:53:06,495 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T20:53:06,495 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-09T20:53:06,495 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T20:53:06,495 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T20:53:06,496 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-09T20:53:06,496 DEBUG [RS:0;00c23a7251f8:39315 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5fb484e4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c23a7251f8/172.17.0.2:0 2024-12-09T20:53:06,496 DEBUG [RS:2;00c23a7251f8:40865 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@31ab286c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c23a7251f8/172.17.0.2:0 2024-12-09T20:53:06,496 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-09T20:53:06,496 DEBUG [RS:1;00c23a7251f8:45461 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3bb57696, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c23a7251f8/172.17.0.2:0 2024-12-09T20:53:06,496 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-09T20:53:06,499 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:06,499 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-09T20:53:06,500 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,512 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:06,513 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:06,514 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-09T20:53:06,515 DEBUG [RS:0;00c23a7251f8:39315 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;00c23a7251f8:39315 2024-12-09T20:53:06,516 DEBUG [RS:1;00c23a7251f8:45461 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;00c23a7251f8:45461 2024-12-09T20:53:06,517 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-09T20:53:06,517 DEBUG [RS:2;00c23a7251f8:40865 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;00c23a7251f8:40865 2024-12-09T20:53:06,518 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-09T20:53:06,519 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T20:53:06,519 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T20:53:06,519 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T20:53:06,519 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T20:53:06,519 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T20:53:06,519 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T20:53:06,519 DEBUG [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T20:53:06,519 DEBUG [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T20:53:06,519 DEBUG [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T20:53:06,522 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c23a7251f8,36171,1733777584597 with port=45461, startcode=1733777585380 2024-12-09T20:53:06,522 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c23a7251f8,36171,1733777584597 with port=39315, startcode=1733777585292 2024-12-09T20:53:06,522 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c23a7251f8,36171,1733777584597 with port=40865, startcode=1733777585425 2024-12-09T20:53:06,525 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-09T20:53:06,525 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-09T20:53:06,527 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.large.0-1733777586527,5,FailOnTimeoutGroup] 2024-12-09T20:53:06,535 DEBUG [RS:2;00c23a7251f8:40865 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T20:53:06,535 DEBUG [RS:1;00c23a7251f8:45461 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T20:53:06,535 DEBUG [RS:0;00c23a7251f8:39315 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T20:53:06,536 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.small.0-1733777586527,5,FailOnTimeoutGroup] 2024-12-09T20:53:06,536 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,536 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-09T20:53:06,538 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,538 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,540 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:32960 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:36667:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:32960 dst: /127.0.0.1:36667 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:06,550 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-09T20:53:06,551 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:06,553 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-09T20:53:06,553 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147 2024-12-09T20:53:06,565 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-09T20:53:06,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-09T20:53:06,579 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:06,579 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:06,584 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:46422 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:34835:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46422 dst: /127.0.0.1:34835 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:06,592 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-09T20:53:06,592 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37887, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T20:53:06,592 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43983, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T20:53:06,592 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56949, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T20:53:06,598 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:06,598 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36171 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c23a7251f8,45461,1733777585380 2024-12-09T20:53:06,599 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:06,601 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36171 {}] master.ServerManager(517): Registering regionserver=00c23a7251f8,45461,1733777585380 2024-12-09T20:53:06,602 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-09T20:53:06,605 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-09T20:53:06,605 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:06,607 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:06,607 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-09T20:53:06,610 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-09T20:53:06,610 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:06,611 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:06,611 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-09T20:53:06,614 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-09T20:53:06,614 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:06,615 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36171 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c23a7251f8,39315,1733777585292 2024-12-09T20:53:06,615 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36171 {}] master.ServerManager(517): Registering regionserver=00c23a7251f8,39315,1733777585292 2024-12-09T20:53:06,615 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:06,615 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-09T20:53:06,618 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-09T20:53:06,618 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:06,619 DEBUG [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147 2024-12-09T20:53:06,619 DEBUG [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33595 2024-12-09T20:53:06,619 DEBUG [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T20:53:06,620 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36171 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c23a7251f8,40865,1733777585425 2024-12-09T20:53:06,620 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36171 {}] master.ServerManager(517): Registering regionserver=00c23a7251f8,40865,1733777585425 2024-12-09T20:53:06,620 DEBUG [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147 2024-12-09T20:53:06,620 DEBUG [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33595 2024-12-09T20:53:06,620 DEBUG [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T20:53:06,621 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:06,621 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-09T20:53:06,624 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740 2024-12-09T20:53:06,625 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740 2024-12-09T20:53:06,625 DEBUG [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147 2024-12-09T20:53:06,625 DEBUG [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33595 2024-12-09T20:53:06,625 DEBUG [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T20:53:06,628 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-09T20:53:06,628 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-09T20:53:06,630 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T20:53:06,632 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-09T20:53:06,638 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T20:53:06,639 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66445332, jitterRate=-0.009887397289276123}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T20:53:06,641 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733777586599Initializing all the Stores at 1733777586601 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777586601Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777586602 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777586602Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777586602Cleaning up temporary data from old regions at 1733777586628 (+26 ms)Region opened successfully at 1733777586641 (+13 ms) 2024-12-09T20:53:06,641 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-09T20:53:06,641 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-09T20:53:06,641 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-09T20:53:06,641 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-09T20:53:06,642 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-09T20:53:06,643 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-09T20:53:06,643 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733777586641Disabling compacts and flushes for region at 1733777586641Disabling writes for close at 1733777586642 (+1 ms)Writing region close event to WAL at 1733777586642Closed at 1733777586643 (+1 ms) 2024-12-09T20:53:06,646 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T20:53:06,646 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-09T20:53:06,651 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-09T20:53:06,658 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T20:53:06,658 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-09T20:53:06,661 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-09T20:53:06,672 DEBUG [RS:0;00c23a7251f8:39315 {}] zookeeper.ZKUtil(111): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c23a7251f8,39315,1733777585292 2024-12-09T20:53:06,672 DEBUG [RS:1;00c23a7251f8:45461 {}] zookeeper.ZKUtil(111): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c23a7251f8,45461,1733777585380 2024-12-09T20:53:06,673 WARN [RS:0;00c23a7251f8:39315 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T20:53:06,673 WARN [RS:1;00c23a7251f8:45461 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T20:53:06,673 INFO [RS:1;00c23a7251f8:45461 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T20:53:06,673 INFO [RS:0;00c23a7251f8:39315 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T20:53:06,673 DEBUG [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,39315,1733777585292 2024-12-09T20:53:06,673 DEBUG [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,45461,1733777585380 2024-12-09T20:53:06,673 DEBUG [RS:2;00c23a7251f8:40865 {}] zookeeper.ZKUtil(111): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c23a7251f8,40865,1733777585425 2024-12-09T20:53:06,673 WARN [RS:2;00c23a7251f8:40865 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T20:53:06,673 INFO [RS:2;00c23a7251f8:40865 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T20:53:06,673 DEBUG [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,40865,1733777585425 2024-12-09T20:53:06,674 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c23a7251f8,40865,1733777585425] 2024-12-09T20:53:06,674 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c23a7251f8,39315,1733777585292] 2024-12-09T20:53:06,674 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c23a7251f8,45461,1733777585380] 2024-12-09T20:53:06,701 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T20:53:06,701 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T20:53:06,701 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T20:53:06,712 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T20:53:06,712 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T20:53:06,713 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T20:53:06,717 INFO [RS:0;00c23a7251f8:39315 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T20:53:06,717 INFO [RS:1;00c23a7251f8:45461 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T20:53:06,717 INFO [RS:2;00c23a7251f8:40865 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T20:53:06,717 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,717 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,717 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,718 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T20:53:06,722 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T20:53:06,722 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T20:53:06,723 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T20:53:06,723 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T20:53:06,723 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T20:53:06,725 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,725 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,725 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,725 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,725 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c23a7251f8:0, corePoolSize=2, maxPoolSize=2 2024-12-09T20:53:06,726 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c23a7251f8:0, corePoolSize=2, maxPoolSize=2 2024-12-09T20:53:06,726 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c23a7251f8:0, corePoolSize=2, maxPoolSize=2 2024-12-09T20:53:06,726 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,726 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:06,727 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:06,727 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:06,727 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:06,727 DEBUG [RS:2;00c23a7251f8:40865 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:06,727 DEBUG [RS:1;00c23a7251f8:45461 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:06,727 DEBUG [RS:0;00c23a7251f8:39315 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:06,732 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,732 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,732 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,732 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,732 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,732 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,732 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,39315,1733777585292-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,40865,1733777585425-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T20:53:06,733 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,45461,1733777585380-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T20:53:06,749 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T20:53:06,749 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T20:53:06,750 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,39315,1733777585292-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,750 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,45461,1733777585380-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,751 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,751 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,751 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.Replication(171): 00c23a7251f8,39315,1733777585292 started 2024-12-09T20:53:06,751 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.Replication(171): 00c23a7251f8,45461,1733777585380 started 2024-12-09T20:53:06,754 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T20:53:06,754 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,40865,1733777585425-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,754 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,754 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.Replication(171): 00c23a7251f8,40865,1733777585425 started 2024-12-09T20:53:06,767 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,767 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1482): Serving as 00c23a7251f8,39315,1733777585292, RpcServer on 00c23a7251f8/172.17.0.2:39315, sessionid=0x1000c5a75aa0001 2024-12-09T20:53:06,768 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T20:53:06,768 DEBUG [RS:0;00c23a7251f8:39315 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c23a7251f8,39315,1733777585292 2024-12-09T20:53:06,768 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,39315,1733777585292' 2024-12-09T20:53:06,769 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T20:53:06,769 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T20:53:06,770 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T20:53:06,770 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T20:53:06,770 DEBUG [RS:0;00c23a7251f8:39315 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c23a7251f8,39315,1733777585292 2024-12-09T20:53:06,770 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,39315,1733777585292' 2024-12-09T20:53:06,771 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T20:53:06,771 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,771 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(1482): Serving as 00c23a7251f8,45461,1733777585380, RpcServer on 00c23a7251f8/172.17.0.2:45461, sessionid=0x1000c5a75aa0002 2024-12-09T20:53:06,771 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T20:53:06,771 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T20:53:06,771 DEBUG [RS:1;00c23a7251f8:45461 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c23a7251f8,45461,1733777585380 2024-12-09T20:53:06,771 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,45461,1733777585380' 2024-12-09T20:53:06,771 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T20:53:06,772 DEBUG [RS:0;00c23a7251f8:39315 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T20:53:06,772 INFO [RS:0;00c23a7251f8:39315 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T20:53:06,772 INFO [RS:0;00c23a7251f8:39315 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T20:53:06,772 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T20:53:06,773 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T20:53:06,773 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T20:53:06,773 DEBUG [RS:1;00c23a7251f8:45461 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c23a7251f8,45461,1733777585380 2024-12-09T20:53:06,773 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,45461,1733777585380' 2024-12-09T20:53:06,773 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T20:53:06,773 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T20:53:06,774 DEBUG [RS:1;00c23a7251f8:45461 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T20:53:06,774 INFO [RS:1;00c23a7251f8:45461 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T20:53:06,774 INFO [RS:1;00c23a7251f8:45461 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T20:53:06,777 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:06,777 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1482): Serving as 00c23a7251f8,40865,1733777585425, RpcServer on 00c23a7251f8/172.17.0.2:40865, sessionid=0x1000c5a75aa0003 2024-12-09T20:53:06,777 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T20:53:06,777 DEBUG [RS:2;00c23a7251f8:40865 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c23a7251f8,40865,1733777585425 2024-12-09T20:53:06,777 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,40865,1733777585425' 2024-12-09T20:53:06,777 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T20:53:06,778 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T20:53:06,779 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T20:53:06,779 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T20:53:06,779 DEBUG [RS:2;00c23a7251f8:40865 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c23a7251f8,40865,1733777585425 2024-12-09T20:53:06,779 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,40865,1733777585425' 2024-12-09T20:53:06,779 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T20:53:06,779 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T20:53:06,780 DEBUG [RS:2;00c23a7251f8:40865 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T20:53:06,780 INFO [RS:2;00c23a7251f8:40865 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T20:53:06,780 INFO [RS:2;00c23a7251f8:40865 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T20:53:06,812 WARN [00c23a7251f8:36171 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-09T20:53:06,879 INFO [RS:0;00c23a7251f8:39315 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-09T20:53:06,879 INFO [RS:1;00c23a7251f8:45461 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-09T20:53:06,881 INFO [RS:2;00c23a7251f8:40865 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-09T20:53:06,884 INFO [RS:1;00c23a7251f8:45461 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C45461%2C1733777585380, suffix=, logDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,45461,1733777585380, archiveDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/oldWALs, maxLogs=32 2024-12-09T20:53:06,884 INFO [RS:0;00c23a7251f8:39315 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C39315%2C1733777585292, suffix=, logDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,39315,1733777585292, archiveDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/oldWALs, maxLogs=32 2024-12-09T20:53:06,884 INFO [RS:2;00c23a7251f8:40865 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C40865%2C1733777585425, suffix=, logDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,40865,1733777585425, archiveDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/oldWALs, maxLogs=32 2024-12-09T20:53:06,899 DEBUG [RS:1;00c23a7251f8:45461 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,45461,1733777585380/00c23a7251f8%2C45461%2C1733777585380.1733777586887, exclude list is [], retry=0 2024-12-09T20:53:06,899 DEBUG [RS:2;00c23a7251f8:40865 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,40865,1733777585425/00c23a7251f8%2C40865%2C1733777585425.1733777586887, exclude list is [], retry=0 2024-12-09T20:53:06,901 DEBUG [RS:0;00c23a7251f8:39315 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,39315,1733777585292/00c23a7251f8%2C39315%2C1733777585292.1733777586887, exclude list is [], retry=0 2024-12-09T20:53:06,904 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34835,DS-f5161ea3-c80e-42e5-8de0-a41d9afc7c11,DISK] 2024-12-09T20:53:06,905 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34835,DS-f5161ea3-c80e-42e5-8de0-a41d9afc7c11,DISK] 2024-12-09T20:53:06,905 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:36667,DS-c2e178b9-c94c-4467-b7d3-5489432fe387,DISK] 2024-12-09T20:53:06,905 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38901,DS-e22c6745-d9a8-48a2-b4b1-814705d31b2e,DISK] 2024-12-09T20:53:06,905 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38901,DS-e22c6745-d9a8-48a2-b4b1-814705d31b2e,DISK] 2024-12-09T20:53:06,906 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:36667,DS-c2e178b9-c94c-4467-b7d3-5489432fe387,DISK] 2024-12-09T20:53:06,930 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38901,DS-e22c6745-d9a8-48a2-b4b1-814705d31b2e,DISK] 2024-12-09T20:53:06,930 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:36667,DS-c2e178b9-c94c-4467-b7d3-5489432fe387,DISK] 2024-12-09T20:53:06,931 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34835,DS-f5161ea3-c80e-42e5-8de0-a41d9afc7c11,DISK] 2024-12-09T20:53:06,939 INFO [RS:2;00c23a7251f8:40865 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,40865,1733777585425/00c23a7251f8%2C40865%2C1733777585425.1733777586887 2024-12-09T20:53:06,940 INFO [RS:1;00c23a7251f8:45461 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,45461,1733777585380/00c23a7251f8%2C45461%2C1733777585380.1733777586887 2024-12-09T20:53:06,941 DEBUG [RS:2;00c23a7251f8:40865 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34993:34993),(127.0.0.1/127.0.0.1:33261:33261),(127.0.0.1/127.0.0.1:43477:43477)] 2024-12-09T20:53:06,942 DEBUG [RS:1;00c23a7251f8:45461 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:43477:43477),(127.0.0.1/127.0.0.1:33261:33261),(127.0.0.1/127.0.0.1:34993:34993)] 2024-12-09T20:53:06,942 INFO [RS:0;00c23a7251f8:39315 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,39315,1733777585292/00c23a7251f8%2C39315%2C1733777585292.1733777586887 2024-12-09T20:53:06,943 DEBUG [RS:0;00c23a7251f8:39315 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:33261:33261),(127.0.0.1/127.0.0.1:34993:34993),(127.0.0.1/127.0.0.1:43477:43477)] 2024-12-09T20:53:07,065 DEBUG [00c23a7251f8:36171 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-09T20:53:07,075 DEBUG [00c23a7251f8:36171 {}] balancer.BalancerClusterState(204): Hosts are {00c23a7251f8=0} racks are {/default-rack=0} 2024-12-09T20:53:07,081 DEBUG [00c23a7251f8:36171 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-09T20:53:07,082 DEBUG [00c23a7251f8:36171 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-09T20:53:07,082 DEBUG [00c23a7251f8:36171 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-09T20:53:07,082 DEBUG [00c23a7251f8:36171 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-09T20:53:07,082 DEBUG [00c23a7251f8:36171 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-09T20:53:07,082 DEBUG [00c23a7251f8:36171 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-09T20:53:07,082 INFO [00c23a7251f8:36171 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-09T20:53:07,082 INFO [00c23a7251f8:36171 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-09T20:53:07,082 INFO [00c23a7251f8:36171 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-09T20:53:07,082 DEBUG [00c23a7251f8:36171 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-09T20:53:07,088 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=00c23a7251f8,39315,1733777585292 2024-12-09T20:53:07,093 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 00c23a7251f8,39315,1733777585292, state=OPENING 2024-12-09T20:53:07,146 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-09T20:53:07,157 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:07,157 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:07,157 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:07,157 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:07,159 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:07,159 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:07,159 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:07,160 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:07,162 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-09T20:53:07,165 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=00c23a7251f8,39315,1733777585292}] 2024-12-09T20:53:07,344 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-09T20:53:07,347 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55981, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-09T20:53:07,358 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-09T20:53:07,358 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T20:53:07,359 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-09T20:53:07,362 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C39315%2C1733777585292.meta, suffix=.meta, logDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,39315,1733777585292, archiveDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/oldWALs, maxLogs=32 2024-12-09T20:53:07,376 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,39315,1733777585292/00c23a7251f8%2C39315%2C1733777585292.meta.1733777587363.meta, exclude list is [], retry=0 2024-12-09T20:53:07,381 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34835,DS-f5161ea3-c80e-42e5-8de0-a41d9afc7c11,DISK] 2024-12-09T20:53:07,381 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38901,DS-e22c6745-d9a8-48a2-b4b1-814705d31b2e,DISK] 2024-12-09T20:53:07,381 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:36667,DS-c2e178b9-c94c-4467-b7d3-5489432fe387,DISK] 2024-12-09T20:53:07,384 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,39315,1733777585292/00c23a7251f8%2C39315%2C1733777585292.meta.1733777587363.meta 2024-12-09T20:53:07,384 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:33261:33261),(127.0.0.1/127.0.0.1:34993:34993),(127.0.0.1/127.0.0.1:43477:43477)] 2024-12-09T20:53:07,384 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-09T20:53:07,386 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-09T20:53:07,388 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-09T20:53:07,391 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-09T20:53:07,412 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-09T20:53:07,413 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:07,413 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-09T20:53:07,413 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-09T20:53:07,417 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-09T20:53:07,419 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-09T20:53:07,419 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:07,420 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:07,420 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-09T20:53:07,421 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-09T20:53:07,422 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:07,422 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:07,422 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-09T20:53:07,424 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-09T20:53:07,424 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:07,425 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:07,425 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-09T20:53:07,426 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-09T20:53:07,426 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:07,427 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:07,427 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-09T20:53:07,428 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740 2024-12-09T20:53:07,431 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740 2024-12-09T20:53:07,434 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-09T20:53:07,434 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-09T20:53:07,434 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T20:53:07,437 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-09T20:53:07,438 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74568122, jitterRate=0.11115160584449768}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T20:53:07,439 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-09T20:53:07,440 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733777587414Writing region info on filesystem at 1733777587414Initializing all the Stores at 1733777587416 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777587417 (+1 ms)Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777587417Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777587417Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777587417Cleaning up temporary data from old regions at 1733777587434 (+17 ms)Running coprocessor post-open hooks at 1733777587439 (+5 ms)Region opened successfully at 1733777587440 (+1 ms) 2024-12-09T20:53:07,450 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733777587337 2024-12-09T20:53:07,462 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-09T20:53:07,463 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-09T20:53:07,464 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=00c23a7251f8,39315,1733777585292 2024-12-09T20:53:07,466 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 00c23a7251f8,39315,1733777585292, state=OPEN 2024-12-09T20:53:07,472 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T20:53:07,472 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T20:53:07,472 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T20:53:07,472 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T20:53:07,472 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:07,472 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:07,472 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:07,472 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:07,473 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=00c23a7251f8,39315,1733777585292 2024-12-09T20:53:07,478 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-09T20:53:07,479 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=00c23a7251f8,39315,1733777585292 in 308 msec 2024-12-09T20:53:07,488 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-09T20:53:07,488 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 830 msec 2024-12-09T20:53:07,490 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T20:53:07,490 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-09T20:53:07,508 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-09T20:53:07,509 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=00c23a7251f8,39315,1733777585292, seqNum=-1] 2024-12-09T20:53:07,525 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T20:53:07,528 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:53695, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T20:53:07,545 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1260 sec 2024-12-09T20:53:07,545 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733777587545, completionTime=-1 2024-12-09T20:53:07,547 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-09T20:53:07,547 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-09T20:53:07,569 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-09T20:53:07,569 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733777647569 2024-12-09T20:53:07,570 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733777707570 2024-12-09T20:53:07,570 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 22 msec 2024-12-09T20:53:07,571 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-09T20:53:07,577 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36171,1733777584597-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:07,578 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36171,1733777584597-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:07,578 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36171,1733777584597-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:07,579 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-00c23a7251f8:36171, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:07,579 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:07,580 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:07,586 DEBUG [master/00c23a7251f8:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-09T20:53:07,611 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.095sec 2024-12-09T20:53:07,612 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-09T20:53:07,613 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-09T20:53:07,614 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-09T20:53:07,614 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-09T20:53:07,614 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-09T20:53:07,615 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36171,1733777584597-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T20:53:07,615 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36171,1733777584597-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-09T20:53:07,620 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-09T20:53:07,621 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-09T20:53:07,621 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36171,1733777584597-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:07,670 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3f922879, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T20:53:07,674 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-09T20:53:07,674 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-09T20:53:07,678 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 00c23a7251f8,36171,-1 for getting cluster id 2024-12-09T20:53:07,680 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-09T20:53:07,687 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '9a9faef4-5ccb-4487-a0e3-789175cc03b6' 2024-12-09T20:53:07,688 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-09T20:53:07,689 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "9a9faef4-5ccb-4487-a0e3-789175cc03b6" 2024-12-09T20:53:07,689 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@300bdfc7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T20:53:07,689 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [00c23a7251f8,36171,-1] 2024-12-09T20:53:07,693 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-09T20:53:07,695 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:07,695 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34486, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-09T20:53:07,698 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@49703e48, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T20:53:07,698 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-09T20:53:07,705 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=00c23a7251f8,39315,1733777585292, seqNum=-1] 2024-12-09T20:53:07,705 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T20:53:07,708 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46298, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T20:53:07,725 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=00c23a7251f8,36171,1733777584597 2024-12-09T20:53:07,728 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-09T20:53:07,733 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 00c23a7251f8,36171,1733777584597 2024-12-09T20:53:07,735 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@2aca6771 2024-12-09T20:53:07,736 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-09T20:53:07,738 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34498, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-09T20:53:07,743 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-09T20:53:07,750 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-09T20:53:07,752 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-09T20:53:07,754 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-09T20:53:07,754 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:07,756 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-09T20:53:07,759 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:07,765 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:07,765 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:07,767 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:33016 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:36667:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33016 dst: /127.0.0.1:36667 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:07,772 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-09T20:53:07,772 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:07,775 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 86e22e17336ef7d8ae8f2697fc87dc4e, NAME => 'TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147 2024-12-09T20:53:07,782 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:07,782 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:07,785 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:33030 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:36667:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33030 dst: /127.0.0.1:36667 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:07,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-09T20:53:07,873 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:08,085 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:08,191 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:08,192 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:08,193 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 86e22e17336ef7d8ae8f2697fc87dc4e, disabling compactions & flushes 2024-12-09T20:53:08,193 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:08,193 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:08,193 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. after waiting 0 ms 2024-12-09T20:53:08,193 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:08,194 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:08,194 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 86e22e17336ef7d8ae8f2697fc87dc4e: Waiting for close lock at 1733777588193Disabling compacts and flushes for region at 1733777588193Disabling writes for close at 1733777588193Writing region close event to WAL at 1733777588194 (+1 ms)Closed at 1733777588194 2024-12-09T20:53:08,199 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-09T20:53:08,205 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733777588199"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733777588199"}]},"ts":"1733777588199"} 2024-12-09T20:53:08,211 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-09T20:53:08,213 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-09T20:53:08,215 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733777588213"}]},"ts":"1733777588213"} 2024-12-09T20:53:08,219 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-09T20:53:08,220 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {00c23a7251f8=0} racks are {/default-rack=0} 2024-12-09T20:53:08,221 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-09T20:53:08,221 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-09T20:53:08,221 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-09T20:53:08,221 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-09T20:53:08,222 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-09T20:53:08,222 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-09T20:53:08,222 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-09T20:53:08,222 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-09T20:53:08,222 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-09T20:53:08,222 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-09T20:53:08,223 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=86e22e17336ef7d8ae8f2697fc87dc4e, ASSIGN}] 2024-12-09T20:53:08,226 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=86e22e17336ef7d8ae8f2697fc87dc4e, ASSIGN 2024-12-09T20:53:08,228 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=86e22e17336ef7d8ae8f2697fc87dc4e, ASSIGN; state=OFFLINE, location=00c23a7251f8,40865,1733777585425; forceNewPlan=false, retain=false 2024-12-09T20:53:08,383 INFO [00c23a7251f8:36171 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-09T20:53:08,384 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=86e22e17336ef7d8ae8f2697fc87dc4e, regionState=OPENING, regionLocation=00c23a7251f8,40865,1733777585425 2024-12-09T20:53:08,388 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=86e22e17336ef7d8ae8f2697fc87dc4e, ASSIGN because future has completed 2024-12-09T20:53:08,390 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 86e22e17336ef7d8ae8f2697fc87dc4e, server=00c23a7251f8,40865,1733777585425}] 2024-12-09T20:53:08,393 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:08,545 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-09T20:53:08,548 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58215, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-09T20:53:08,556 INFO [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:08,557 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 86e22e17336ef7d8ae8f2697fc87dc4e, NAME => 'TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e.', STARTKEY => '', ENDKEY => ''} 2024-12-09T20:53:08,557 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,557 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:08,557 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,557 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,559 INFO [StoreOpener-86e22e17336ef7d8ae8f2697fc87dc4e-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,561 INFO [StoreOpener-86e22e17336ef7d8ae8f2697fc87dc4e-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 86e22e17336ef7d8ae8f2697fc87dc4e columnFamilyName cf 2024-12-09T20:53:08,561 DEBUG [StoreOpener-86e22e17336ef7d8ae8f2697fc87dc4e-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:08,562 INFO [StoreOpener-86e22e17336ef7d8ae8f2697fc87dc4e-1 {}] regionserver.HStore(327): Store=86e22e17336ef7d8ae8f2697fc87dc4e/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T20:53:08,562 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,564 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,564 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,565 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,565 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,568 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,572 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T20:53:08,573 INFO [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 86e22e17336ef7d8ae8f2697fc87dc4e; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60140152, jitterRate=-0.10384190082550049}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-09T20:53:08,573 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:08,574 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 86e22e17336ef7d8ae8f2697fc87dc4e: Running coprocessor pre-open hook at 1733777588557Writing region info on filesystem at 1733777588557Initializing all the Stores at 1733777588559 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777588559Cleaning up temporary data from old regions at 1733777588565 (+6 ms)Running coprocessor post-open hooks at 1733777588573 (+8 ms)Region opened successfully at 1733777588574 (+1 ms) 2024-12-09T20:53:08,576 INFO [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e., pid=6, masterSystemTime=1733777588544 2024-12-09T20:53:08,579 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:08,579 INFO [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:08,580 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=86e22e17336ef7d8ae8f2697fc87dc4e, regionState=OPEN, openSeqNum=2, regionLocation=00c23a7251f8,40865,1733777585425 2024-12-09T20:53:08,584 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 86e22e17336ef7d8ae8f2697fc87dc4e, server=00c23a7251f8,40865,1733777585425 because future has completed 2024-12-09T20:53:08,590 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-09T20:53:08,591 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 86e22e17336ef7d8ae8f2697fc87dc4e, server=00c23a7251f8,40865,1733777585425 in 197 msec 2024-12-09T20:53:08,594 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-09T20:53:08,594 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=86e22e17336ef7d8ae8f2697fc87dc4e, ASSIGN in 367 msec 2024-12-09T20:53:08,595 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-09T20:53:08,595 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733777588595"}]},"ts":"1733777588595"} 2024-12-09T20:53:08,598 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-09T20:53:08,600 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-09T20:53:08,603 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 854 msec 2024-12-09T20:53:08,903 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:08,905 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-09T20:53:08,905 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-09T20:53:08,907 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-09T20:53:08,915 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-09T20:53:08,915 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-09T20:53:08,916 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-09T20:53:08,927 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e., hostname=00c23a7251f8,40865,1733777585425, seqNum=2] 2024-12-09T20:53:08,929 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T20:53:08,930 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50526, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T20:53:08,939 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-09T20:53:08,943 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-09T20:53:08,945 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-09T20:53:08,946 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T20:53:08,947 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-09T20:53:08,948 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-09T20:53:09,053 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T20:53:09,117 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=40865 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-09T20:53:09,117 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:09,120 INFO [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 86e22e17336ef7d8ae8f2697fc87dc4e 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-09T20:53:09,167 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e/.tmp/cf/4a244747e7a240f9baf0e79e2f7436ff is 36, key is row/cf:cq/1733777588931/Put/seqid=0 2024-12-09T20:53:09,173 WARN [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,173 WARN [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,176 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1327908529_22 at /127.0.0.1:33048 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:36667:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33048 dst: /127.0.0.1:36667 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:09,181 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-09T20:53:09,182 WARN [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:09,182 INFO [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e/.tmp/cf/4a244747e7a240f9baf0e79e2f7436ff 2024-12-09T20:53:09,219 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e/.tmp/cf/4a244747e7a240f9baf0e79e2f7436ff as hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e/cf/4a244747e7a240f9baf0e79e2f7436ff 2024-12-09T20:53:09,229 INFO [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e/cf/4a244747e7a240f9baf0e79e2f7436ff, entries=1, sequenceid=5, filesize=4.7 K 2024-12-09T20:53:09,236 INFO [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 86e22e17336ef7d8ae8f2697fc87dc4e in 114ms, sequenceid=5, compaction requested=false 2024-12-09T20:53:09,237 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-09T20:53:09,239 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 86e22e17336ef7d8ae8f2697fc87dc4e: 2024-12-09T20:53:09,239 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:09,241 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-09T20:53:09,243 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-09T20:53:09,248 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-09T20:53:09,249 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 297 msec 2024-12-09T20:53:09,253 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 310 msec 2024-12-09T20:53:09,263 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36171 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T20:53:09,263 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-09T20:53:09,275 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-09T20:53:09,276 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-09T20:53:09,276 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:09,280 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,280 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,280 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-09T20:53:09,281 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-09T20:53:09,281 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1053966274, stopped=false 2024-12-09T20:53:09,281 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=00c23a7251f8,36171,1733777584597 2024-12-09T20:53:09,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:09,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:09,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:09,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:09,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:09,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:09,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:09,346 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:09,346 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-09T20:53:09,346 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-09T20:53:09,347 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:09,347 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:09,347 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:09,347 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,347 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:09,347 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c23a7251f8,39315,1733777585292' ***** 2024-12-09T20:53:09,347 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T20:53:09,347 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:09,348 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c23a7251f8,45461,1733777585380' ***** 2024-12-09T20:53:09,348 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T20:53:09,348 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c23a7251f8,40865,1733777585425' ***** 2024-12-09T20:53:09,348 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T20:53:09,348 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T20:53:09,348 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T20:53:09,348 INFO [RS:1;00c23a7251f8:45461 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T20:53:09,348 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T20:53:09,348 INFO [RS:0;00c23a7251f8:39315 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T20:53:09,348 INFO [RS:1;00c23a7251f8:45461 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T20:53:09,349 INFO [RS:0;00c23a7251f8:39315 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T20:53:09,349 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T20:53:09,349 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(959): stopping server 00c23a7251f8,45461,1733777585380 2024-12-09T20:53:09,348 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T20:53:09,349 INFO [RS:2;00c23a7251f8:40865 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T20:53:09,349 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(959): stopping server 00c23a7251f8,39315,1733777585292 2024-12-09T20:53:09,349 INFO [RS:2;00c23a7251f8:40865 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T20:53:09,349 INFO [RS:0;00c23a7251f8:39315 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T20:53:09,349 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T20:53:09,349 INFO [RS:0;00c23a7251f8:39315 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;00c23a7251f8:39315. 2024-12-09T20:53:09,349 INFO [RS:1;00c23a7251f8:45461 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T20:53:09,349 DEBUG [RS:0;00c23a7251f8:39315 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:09,349 INFO [RS:1;00c23a7251f8:45461 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;00c23a7251f8:45461. 2024-12-09T20:53:09,349 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(3091): Received CLOSE for 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:09,349 DEBUG [RS:0;00c23a7251f8:39315 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,349 DEBUG [RS:1;00c23a7251f8:45461 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:09,349 DEBUG [RS:1;00c23a7251f8:45461 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,349 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T20:53:09,349 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T20:53:09,349 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(976): stopping server 00c23a7251f8,45461,1733777585380; all regions closed. 2024-12-09T20:53:09,349 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T20:53:09,350 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-09T20:53:09,350 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(959): stopping server 00c23a7251f8,40865,1733777585425 2024-12-09T20:53:09,350 INFO [RS:2;00c23a7251f8:40865 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T20:53:09,350 INFO [RS:2;00c23a7251f8:40865 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;00c23a7251f8:40865. 2024-12-09T20:53:09,350 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-09T20:53:09,350 DEBUG [RS:2;00c23a7251f8:40865 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:09,350 DEBUG [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-09T20:53:09,350 DEBUG [RS:2;00c23a7251f8:40865 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,350 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-09T20:53:09,350 DEBUG [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1325): Online Regions={86e22e17336ef7d8ae8f2697fc87dc4e=TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e.} 2024-12-09T20:53:09,350 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 86e22e17336ef7d8ae8f2697fc87dc4e, disabling compactions & flushes 2024-12-09T20:53:09,350 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-09T20:53:09,350 DEBUG [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-09T20:53:09,350 INFO [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:09,350 DEBUG [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1351): Waiting on 86e22e17336ef7d8ae8f2697fc87dc4e 2024-12-09T20:53:09,350 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:09,350 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-09T20:53:09,350 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. after waiting 0 ms 2024-12-09T20:53:09,351 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-09T20:53:09,351 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:09,351 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-09T20:53:09,351 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-09T20:53:09,351 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-09T20:53:09,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_1073741826_1016 (size=93) 2024-12-09T20:53:09,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_1073741826_1016 (size=93) 2024-12-09T20:53:09,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_1073741826_1016 (size=93) 2024-12-09T20:53:09,365 DEBUG [RS:1;00c23a7251f8:45461 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/oldWALs 2024-12-09T20:53:09,366 INFO [RS:1;00c23a7251f8:45461 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 00c23a7251f8%2C45461%2C1733777585380:(num 1733777586887) 2024-12-09T20:53:09,366 DEBUG [RS:1;00c23a7251f8:45461 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,366 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:09,366 INFO [RS:1;00c23a7251f8:45461 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T20:53:09,366 INFO [RS:1;00c23a7251f8:45461 {}] hbase.ChoreService(370): Chore service for: regionserver/00c23a7251f8:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-09T20:53:09,367 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T20:53:09,367 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T20:53:09,367 INFO [regionserver/00c23a7251f8:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T20:53:09,367 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T20:53:09,367 INFO [RS:1;00c23a7251f8:45461 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T20:53:09,367 INFO [RS:1;00c23a7251f8:45461 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:45461 2024-12-09T20:53:09,377 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/default/TestHBaseWalOnEC/86e22e17336ef7d8ae8f2697fc87dc4e/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-09T20:53:09,377 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T20:53:09,377 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c23a7251f8,45461,1733777585380 2024-12-09T20:53:09,377 INFO [RS:1;00c23a7251f8:45461 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T20:53:09,379 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c23a7251f8,45461,1733777585380] 2024-12-09T20:53:09,379 INFO [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:09,379 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 86e22e17336ef7d8ae8f2697fc87dc4e: Waiting for close lock at 1733777589350Running coprocessor pre-close hooks at 1733777589350Disabling compacts and flushes for region at 1733777589350Disabling writes for close at 1733777589351 (+1 ms)Writing region close event to WAL at 1733777589352 (+1 ms)Running coprocessor post-close hooks at 1733777589378 (+26 ms)Closed at 1733777589379 (+1 ms) 2024-12-09T20:53:09,380 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e. 2024-12-09T20:53:09,390 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/info/af303afb37f042bba2e090ec65291894 is 153, key is TestHBaseWalOnEC,,1733777587739.86e22e17336ef7d8ae8f2697fc87dc4e./info:regioninfo/1733777588580/Put/seqid=0 2024-12-09T20:53:09,393 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,393 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,397 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_178934654_22 at /127.0.0.1:35174 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:38901:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35174 dst: /127.0.0.1:38901 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:09,398 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c23a7251f8,45461,1733777585380 already deleted, retry=false 2024-12-09T20:53:09,398 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c23a7251f8,45461,1733777585380 expired; onlineServers=2 2024-12-09T20:53:09,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-09T20:53:09,401 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:09,401 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/info/af303afb37f042bba2e090ec65291894 2024-12-09T20:53:09,425 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/ns/d72c92229eef4b85b6ae1ff4bec9fd6a is 43, key is default/ns:d/1733777587531/Put/seqid=0 2024-12-09T20:53:09,427 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,428 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,431 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_178934654_22 at /127.0.0.1:35210 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:38901:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35210 dst: /127.0.0.1:38901 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:09,435 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-09T20:53:09,436 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:09,436 INFO [regionserver/00c23a7251f8:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:09,436 INFO [regionserver/00c23a7251f8:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:09,436 INFO [regionserver/00c23a7251f8:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:09,436 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/ns/d72c92229eef4b85b6ae1ff4bec9fd6a 2024-12-09T20:53:09,461 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/table/94d882cecb174d99aff4e72e68c38500 is 52, key is TestHBaseWalOnEC/table:state/1733777588595/Put/seqid=0 2024-12-09T20:53:09,463 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,463 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,467 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_178934654_22 at /127.0.0.1:35230 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:38901:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35230 dst: /127.0.0.1:38901 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:09,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-09T20:53:09,472 WARN [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:09,472 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/table/94d882cecb174d99aff4e72e68c38500 2024-12-09T20:53:09,482 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/info/af303afb37f042bba2e090ec65291894 as hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/info/af303afb37f042bba2e090ec65291894 2024-12-09T20:53:09,488 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:09,488 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45461-0x1000c5a75aa0002, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:09,489 INFO [RS:1;00c23a7251f8:45461 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T20:53:09,489 INFO [RS:1;00c23a7251f8:45461 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c23a7251f8,45461,1733777585380; zookeeper connection closed. 2024-12-09T20:53:09,489 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4636a574 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4636a574 2024-12-09T20:53:09,492 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/info/af303afb37f042bba2e090ec65291894, entries=10, sequenceid=11, filesize=6.5 K 2024-12-09T20:53:09,494 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/ns/d72c92229eef4b85b6ae1ff4bec9fd6a as hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/ns/d72c92229eef4b85b6ae1ff4bec9fd6a 2024-12-09T20:53:09,503 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/ns/d72c92229eef4b85b6ae1ff4bec9fd6a, entries=2, sequenceid=11, filesize=5.0 K 2024-12-09T20:53:09,505 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/.tmp/table/94d882cecb174d99aff4e72e68c38500 as hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/table/94d882cecb174d99aff4e72e68c38500 2024-12-09T20:53:09,514 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/table/94d882cecb174d99aff4e72e68c38500, entries=2, sequenceid=11, filesize=5.1 K 2024-12-09T20:53:09,516 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 165ms, sequenceid=11, compaction requested=false 2024-12-09T20:53:09,516 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-09T20:53:09,522 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775773_1004 (size=42) 2024-12-09T20:53:09,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775772_1004 (size=42) 2024-12-09T20:53:09,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775725_1010 (size=34) 2024-12-09T20:53:09,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775741_1008 (size=1189) 2024-12-09T20:53:09,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_-9223372036854775724_1010 (size=34) 2024-12-09T20:53:09,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775693_1015 (size=32) 2024-12-09T20:53:09,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775740_1008 (size=1189) 2024-12-09T20:53:09,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775692_1015 (size=32) 2024-12-09T20:53:09,541 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-09T20:53:09,542 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-09T20:53:09,542 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-09T20:53:09,542 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733777589350Running coprocessor pre-close hooks at 1733777589350Disabling compacts and flushes for region at 1733777589350Disabling writes for close at 1733777589351 (+1 ms)Obtaining lock to block concurrent updates at 1733777589351Preparing flush snapshotting stores in 1588230740 at 1733777589351Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733777589352 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733777589355 (+3 ms)Flushing 1588230740/info: creating writer at 1733777589355Flushing 1588230740/info: appending metadata at 1733777589387 (+32 ms)Flushing 1588230740/info: closing flushed file at 1733777589388 (+1 ms)Flushing 1588230740/ns: creating writer at 1733777589410 (+22 ms)Flushing 1588230740/ns: appending metadata at 1733777589424 (+14 ms)Flushing 1588230740/ns: closing flushed file at 1733777589424Flushing 1588230740/table: creating writer at 1733777589444 (+20 ms)Flushing 1588230740/table: appending metadata at 1733777589459 (+15 ms)Flushing 1588230740/table: closing flushed file at 1733777589460 (+1 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1271e2de: reopening flushed file at 1733777589481 (+21 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7a869ce5: reopening flushed file at 1733777589492 (+11 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@13d81317: reopening flushed file at 1733777589504 (+12 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 165ms, sequenceid=11, compaction requested=false at 1733777589516 (+12 ms)Writing region close event to WAL at 1733777589528 (+12 ms)Running coprocessor post-close hooks at 1733777589542 (+14 ms)Closed at 1733777589542 2024-12-09T20:53:09,543 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-09T20:53:09,551 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(976): stopping server 00c23a7251f8,39315,1733777585292; all regions closed. 2024-12-09T20:53:09,551 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(976): stopping server 00c23a7251f8,40865,1733777585425; all regions closed. 2024-12-09T20:53:09,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_1073741829_1019 (size=2751) 2024-12-09T20:53:09,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_1073741829_1019 (size=2751) 2024-12-09T20:53:09,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_1073741829_1019 (size=2751) 2024-12-09T20:53:09,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_1073741827_1017 (size=1298) 2024-12-09T20:53:09,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_1073741827_1017 (size=1298) 2024-12-09T20:53:09,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_1073741827_1017 (size=1298) 2024-12-09T20:53:09,557 DEBUG [RS:2;00c23a7251f8:40865 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/oldWALs 2024-12-09T20:53:09,557 INFO [RS:2;00c23a7251f8:40865 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 00c23a7251f8%2C40865%2C1733777585425:(num 1733777586887) 2024-12-09T20:53:09,557 DEBUG [RS:0;00c23a7251f8:39315 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/oldWALs 2024-12-09T20:53:09,557 DEBUG [RS:2;00c23a7251f8:40865 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,558 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:09,558 INFO [RS:0;00c23a7251f8:39315 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 00c23a7251f8%2C39315%2C1733777585292.meta:.meta(num 1733777587363) 2024-12-09T20:53:09,558 INFO [RS:2;00c23a7251f8:40865 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T20:53:09,558 INFO [RS:2;00c23a7251f8:40865 {}] hbase.ChoreService(370): Chore service for: regionserver/00c23a7251f8:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-09T20:53:09,558 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T20:53:09,558 INFO [regionserver/00c23a7251f8:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T20:53:09,558 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T20:53:09,558 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T20:53:09,558 INFO [RS:2;00c23a7251f8:40865 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T20:53:09,558 INFO [RS:2;00c23a7251f8:40865 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:40865 2024-12-09T20:53:09,562 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_1073741828_1018 (size=93) 2024-12-09T20:53:09,562 WARN [Close-WAL-Writer-0 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(650): complete file /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/WALs/00c23a7251f8,39315,1733777585292/00c23a7251f8%2C39315%2C1733777585292.1733777586887 not finished, retry = 0 2024-12-09T20:53:09,562 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_1073741828_1018 (size=93) 2024-12-09T20:53:09,563 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_1073741828_1018 (size=93) 2024-12-09T20:53:09,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_-9223372036854775757_1006 (size=196) 2024-12-09T20:53:09,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775756_1006 (size=196) 2024-12-09T20:53:09,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775708_1013 (size=1321) 2024-12-09T20:53:09,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_-9223372036854775709_1013 (size=1321) 2024-12-09T20:53:09,588 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T20:53:09,588 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c23a7251f8,40865,1733777585425 2024-12-09T20:53:09,588 INFO [RS:2;00c23a7251f8:40865 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T20:53:09,598 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c23a7251f8,40865,1733777585425] 2024-12-09T20:53:09,609 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c23a7251f8,40865,1733777585425 already deleted, retry=false 2024-12-09T20:53:09,609 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c23a7251f8,40865,1733777585425 expired; onlineServers=1 2024-12-09T20:53:09,673 DEBUG [RS:0;00c23a7251f8:39315 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/oldWALs 2024-12-09T20:53:09,673 INFO [RS:0;00c23a7251f8:39315 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 00c23a7251f8%2C39315%2C1733777585292:(num 1733777586887) 2024-12-09T20:53:09,673 DEBUG [RS:0;00c23a7251f8:39315 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:09,673 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:09,673 INFO [RS:0;00c23a7251f8:39315 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T20:53:09,673 INFO [RS:0;00c23a7251f8:39315 {}] hbase.ChoreService(370): Chore service for: regionserver/00c23a7251f8:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-09T20:53:09,674 INFO [RS:0;00c23a7251f8:39315 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T20:53:09,674 INFO [regionserver/00c23a7251f8:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T20:53:09,674 INFO [RS:0;00c23a7251f8:39315 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39315 2024-12-09T20:53:09,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c23a7251f8,39315,1733777585292 2024-12-09T20:53:09,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T20:53:09,682 INFO [RS:0;00c23a7251f8:39315 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T20:53:09,693 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c23a7251f8,39315,1733777585292] 2024-12-09T20:53:09,698 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:09,698 INFO [RS:2;00c23a7251f8:40865 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T20:53:09,698 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40865-0x1000c5a75aa0003, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:09,699 INFO [RS:2;00c23a7251f8:40865 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c23a7251f8,40865,1733777585425; zookeeper connection closed. 2024-12-09T20:53:09,699 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@2f2cee39 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@2f2cee39 2024-12-09T20:53:09,703 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c23a7251f8,39315,1733777585292 already deleted, retry=false 2024-12-09T20:53:09,703 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c23a7251f8,39315,1733777585292 expired; onlineServers=0 2024-12-09T20:53:09,703 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '00c23a7251f8,36171,1733777584597' ***** 2024-12-09T20:53:09,703 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-09T20:53:09,704 INFO [M:0;00c23a7251f8:36171 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T20:53:09,704 INFO [M:0;00c23a7251f8:36171 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T20:53:09,704 DEBUG [M:0;00c23a7251f8:36171 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-09T20:53:09,704 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-09T20:53:09,704 DEBUG [M:0;00c23a7251f8:36171 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-09T20:53:09,704 DEBUG [master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.large.0-1733777586527 {}] cleaner.HFileCleaner(306): Exit Thread[master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.large.0-1733777586527,5,FailOnTimeoutGroup] 2024-12-09T20:53:09,704 DEBUG [master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.small.0-1733777586527 {}] cleaner.HFileCleaner(306): Exit Thread[master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.small.0-1733777586527,5,FailOnTimeoutGroup] 2024-12-09T20:53:09,704 INFO [M:0;00c23a7251f8:36171 {}] hbase.ChoreService(370): Chore service for: master/00c23a7251f8:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-09T20:53:09,704 INFO [M:0;00c23a7251f8:36171 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T20:53:09,705 DEBUG [M:0;00c23a7251f8:36171 {}] master.HMaster(1795): Stopping service threads 2024-12-09T20:53:09,705 INFO [M:0;00c23a7251f8:36171 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-09T20:53:09,705 INFO [M:0;00c23a7251f8:36171 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-09T20:53:09,705 INFO [M:0;00c23a7251f8:36171 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-09T20:53:09,705 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-09T20:53:09,714 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-09T20:53:09,714 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:09,714 DEBUG [M:0;00c23a7251f8:36171 {}] zookeeper.ZKUtil(347): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-09T20:53:09,714 WARN [M:0;00c23a7251f8:36171 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-09T20:53:09,716 INFO [M:0;00c23a7251f8:36171 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/.lastflushedseqids 2024-12-09T20:53:09,726 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,726 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,728 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:33138 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:36667:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33138 dst: /127.0.0.1:36667 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:09,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-09T20:53:09,733 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:09,733 INFO [M:0;00c23a7251f8:36171 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-09T20:53:09,733 INFO [M:0;00c23a7251f8:36171 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-09T20:53:09,733 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-09T20:53:09,733 INFO [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:09,734 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:09,734 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-09T20:53:09,734 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:09,734 INFO [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.85 KB heapSize=34.13 KB 2024-12-09T20:53:09,749 DEBUG [M:0;00c23a7251f8:36171 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/528e383609034b8d96937a385c8bae7a is 82, key is hbase:meta,,1/info:regioninfo/1733777587463/Put/seqid=0 2024-12-09T20:53:09,751 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,751 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,753 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:35292 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:38901:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35292 dst: /127.0.0.1:38901 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:09,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-09T20:53:09,757 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:09,757 INFO [M:0;00c23a7251f8:36171 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/528e383609034b8d96937a385c8bae7a 2024-12-09T20:53:09,780 DEBUG [M:0;00c23a7251f8:36171 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/45ecee8a64a74e4491b821784c1266c5 is 749, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733777588602/Put/seqid=0 2024-12-09T20:53:09,783 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,783 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,786 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:33164 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:36667:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33164 dst: /127.0.0.1:36667 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:09,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_-9223372036854775552_1037 (size=6441) 2024-12-09T20:53:09,790 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:09,790 INFO [M:0;00c23a7251f8:36171 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.17 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/45ecee8a64a74e4491b821784c1266c5 2024-12-09T20:53:09,793 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:09,793 INFO [RS:0;00c23a7251f8:39315 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T20:53:09,793 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39315-0x1000c5a75aa0001, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:09,793 INFO [RS:0;00c23a7251f8:39315 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c23a7251f8,39315,1733777585292; zookeeper connection closed. 2024-12-09T20:53:09,793 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@332a229 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@332a229 2024-12-09T20:53:09,794 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-09T20:53:09,811 DEBUG [M:0;00c23a7251f8:36171 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/a53f961611d9451094f3f81a5b22f170 is 69, key is 00c23a7251f8,39315,1733777585292/rs:state/1733777586615/Put/seqid=0 2024-12-09T20:53:09,813 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,813 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T20:53:09,815 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-605698089_22 at /127.0.0.1:46520 [Receiving block BP-1840173676-172.17.0.2-1733777580185:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:34835:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46520 dst: /127.0.0.1:34835 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T20:53:09,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-09T20:53:09,819 WARN [M:0;00c23a7251f8:36171 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T20:53:09,819 INFO [M:0;00c23a7251f8:36171 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/a53f961611d9451094f3f81a5b22f170 2024-12-09T20:53:09,828 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/528e383609034b8d96937a385c8bae7a as hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/528e383609034b8d96937a385c8bae7a 2024-12-09T20:53:09,837 INFO [M:0;00c23a7251f8:36171 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/528e383609034b8d96937a385c8bae7a, entries=8, sequenceid=72, filesize=5.5 K 2024-12-09T20:53:09,839 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/45ecee8a64a74e4491b821784c1266c5 as hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/45ecee8a64a74e4491b821784c1266c5 2024-12-09T20:53:09,847 INFO [M:0;00c23a7251f8:36171 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/45ecee8a64a74e4491b821784c1266c5, entries=8, sequenceid=72, filesize=6.3 K 2024-12-09T20:53:09,848 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/a53f961611d9451094f3f81a5b22f170 as hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/a53f961611d9451094f3f81a5b22f170 2024-12-09T20:53:09,855 INFO [M:0;00c23a7251f8:36171 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/a53f961611d9451094f3f81a5b22f170, entries=3, sequenceid=72, filesize=5.2 K 2024-12-09T20:53:09,856 INFO [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 122ms, sequenceid=72, compaction requested=false 2024-12-09T20:53:09,858 INFO [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:09,858 DEBUG [M:0;00c23a7251f8:36171 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733777589733Disabling compacts and flushes for region at 1733777589733Disabling writes for close at 1733777589734 (+1 ms)Obtaining lock to block concurrent updates at 1733777589734Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733777589734Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27492, getHeapSize=34888, getOffHeapSize=0, getCellsCount=85 at 1733777589734Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733777589735 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733777589735Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733777589748 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733777589748Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733777589766 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733777589780 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733777589780Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733777589797 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733777589811 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733777589811Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@773ebafa: reopening flushed file at 1733777589827 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@50733844: reopening flushed file at 1733777589837 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7d2bba08: reopening flushed file at 1733777589847 (+10 ms)Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 122ms, sequenceid=72, compaction requested=false at 1733777589856 (+9 ms)Writing region close event to WAL at 1733777589858 (+2 ms)Closed at 1733777589858 2024-12-09T20:53:09,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36667 is added to blk_1073741825_1011 (size=32695) 2024-12-09T20:53:09,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38901 is added to blk_1073741825_1011 (size=32695) 2024-12-09T20:53:09,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34835 is added to blk_1073741825_1011 (size=32695) 2024-12-09T20:53:09,862 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T20:53:09,862 INFO [M:0;00c23a7251f8:36171 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-09T20:53:09,863 INFO [M:0;00c23a7251f8:36171 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36171 2024-12-09T20:53:09,863 INFO [M:0;00c23a7251f8:36171 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T20:53:09,972 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:09,972 INFO [M:0;00c23a7251f8:36171 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T20:53:09,973 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36171-0x1000c5a75aa0000, quorum=127.0.0.1:53435, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:10,013 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3114ae69{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:10,016 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3c70a874{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T20:53:10,016 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T20:53:10,017 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5822645a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T20:53:10,017 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@16cd567f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir/,STOPPED} 2024-12-09T20:53:10,021 WARN [BP-1840173676-172.17.0.2-1733777580185 heartbeating to localhost/127.0.0.1:33595 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T20:53:10,021 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T20:53:10,021 WARN [BP-1840173676-172.17.0.2-1733777580185 heartbeating to localhost/127.0.0.1:33595 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1840173676-172.17.0.2-1733777580185 (Datanode Uuid 40225027-bbd4-49de-98d7-4b182bc0f27b) service to localhost/127.0.0.1:33595 2024-12-09T20:53:10,021 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T20:53:10,022 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data5/current/BP-1840173676-172.17.0.2-1733777580185 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:10,023 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data6/current/BP-1840173676-172.17.0.2-1733777580185 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:10,023 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T20:53:10,029 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@353955e9{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:10,030 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11738cd8{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T20:53:10,030 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T20:53:10,030 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@40eb7053{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T20:53:10,030 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@510fec09{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir/,STOPPED} 2024-12-09T20:53:10,031 WARN [BP-1840173676-172.17.0.2-1733777580185 heartbeating to localhost/127.0.0.1:33595 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T20:53:10,031 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T20:53:10,031 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T20:53:10,031 WARN [BP-1840173676-172.17.0.2-1733777580185 heartbeating to localhost/127.0.0.1:33595 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1840173676-172.17.0.2-1733777580185 (Datanode Uuid 0af9d13e-06ac-4a26-9f56-90d344b0197a) service to localhost/127.0.0.1:33595 2024-12-09T20:53:10,032 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data3/current/BP-1840173676-172.17.0.2-1733777580185 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:10,032 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data4/current/BP-1840173676-172.17.0.2-1733777580185 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:10,032 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T20:53:10,034 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1b97a472{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:10,034 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3722a29b{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T20:53:10,034 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T20:53:10,034 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@69893329{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T20:53:10,035 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3a5de9e4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir/,STOPPED} 2024-12-09T20:53:10,036 WARN [BP-1840173676-172.17.0.2-1733777580185 heartbeating to localhost/127.0.0.1:33595 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T20:53:10,036 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T20:53:10,036 WARN [BP-1840173676-172.17.0.2-1733777580185 heartbeating to localhost/127.0.0.1:33595 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1840173676-172.17.0.2-1733777580185 (Datanode Uuid 9841445a-8140-4932-8cc1-7dc15f099dc2) service to localhost/127.0.0.1:33595 2024-12-09T20:53:10,036 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T20:53:10,036 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data1/current/BP-1840173676-172.17.0.2-1733777580185 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:10,036 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/cluster_87ba0576-2ff1-c290-5beb-79c4a64a9416/data/data2/current/BP-1840173676-172.17.0.2-1733777580185 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:10,037 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T20:53:10,042 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@62d6efd9{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-09T20:53:10,043 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@353d35a1{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T20:53:10,043 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T20:53:10,043 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@ce709a8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T20:53:10,043 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@760c69c0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir/,STOPPED} 2024-12-09T20:53:10,050 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-09T20:53:10,078 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-09T20:53:10,083 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=91 (was 161), OpenFileDescriptor=439 (was 393) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=217 (was 239), ProcessCount=11 (was 11), AvailableMemoryMB=3891 (was 4192) 2024-12-09T20:53:10,089 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=91, OpenFileDescriptor=439, MaxFileDescriptor=1048576, SystemLoadAverage=217, ProcessCount=11, AvailableMemoryMB=3891 2024-12-09T20:53:10,089 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-09T20:53:10,089 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.log.dir so I do NOT create it in target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4 2024-12-09T20:53:10,089 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/3afc2c52-4c8e-dd81-bbde-65d87164ae33/hadoop.tmp.dir so I do NOT create it in target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4 2024-12-09T20:53:10,089 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61, deleteOnExit=true 2024-12-09T20:53:10,089 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-09T20:53:10,090 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/test.cache.data in system properties and HBase conf 2024-12-09T20:53:10,090 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.tmp.dir in system properties and HBase conf 2024-12-09T20:53:10,090 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir in system properties and HBase conf 2024-12-09T20:53:10,090 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-09T20:53:10,090 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-09T20:53:10,090 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-09T20:53:10,090 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-09T20:53:10,090 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/nfs.dump.dir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/java.io.tmpdir in system properties and HBase conf 2024-12-09T20:53:10,091 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-09T20:53:10,092 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-09T20:53:10,092 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-09T20:53:10,402 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:10,406 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T20:53:10,409 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T20:53:10,409 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T20:53:10,409 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-09T20:53:10,410 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:10,410 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1aa34083{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir/,AVAILABLE} 2024-12-09T20:53:10,410 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@59bbe271{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T20:53:10,499 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@563c957f{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/java.io.tmpdir/jetty-localhost-40167-hadoop-hdfs-3_4_1-tests_jar-_-any-3279703232247086067/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-09T20:53:10,500 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@cbd9f23{HTTP/1.1, (http/1.1)}{localhost:40167} 2024-12-09T20:53:10,500 INFO [Time-limited test {}] server.Server(415): Started @11971ms 2024-12-09T20:53:10,780 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:10,783 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T20:53:10,784 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T20:53:10,784 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T20:53:10,784 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-09T20:53:10,785 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6decf963{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir/,AVAILABLE} 2024-12-09T20:53:10,785 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@46f2e60d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T20:53:10,874 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@765c7210{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/java.io.tmpdir/jetty-localhost-37257-hadoop-hdfs-3_4_1-tests_jar-_-any-6010047846795843701/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:10,874 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3e498d5c{HTTP/1.1, (http/1.1)}{localhost:37257} 2024-12-09T20:53:10,874 INFO [Time-limited test {}] server.Server(415): Started @12345ms 2024-12-09T20:53:10,875 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T20:53:10,907 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:10,910 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T20:53:10,912 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T20:53:10,912 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T20:53:10,912 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-09T20:53:10,912 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7febc9c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir/,AVAILABLE} 2024-12-09T20:53:10,913 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6f5c60f4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T20:53:11,002 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4be50faa{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/java.io.tmpdir/jetty-localhost-46843-hadoop-hdfs-3_4_1-tests_jar-_-any-573984058966911878/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:11,003 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@6625a4f2{HTTP/1.1, (http/1.1)}{localhost:46843} 2024-12-09T20:53:11,003 INFO [Time-limited test {}] server.Server(415): Started @12473ms 2024-12-09T20:53:11,004 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T20:53:11,029 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T20:53:11,033 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T20:53:11,034 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T20:53:11,034 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T20:53:11,034 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-09T20:53:11,034 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@39179133{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir/,AVAILABLE} 2024-12-09T20:53:11,035 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@60d8940e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T20:53:11,127 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@733029a3{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/java.io.tmpdir/jetty-localhost-36293-hadoop-hdfs-3_4_1-tests_jar-_-any-11466111572236820477/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:11,127 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@a87cd5{HTTP/1.1, (http/1.1)}{localhost:36293} 2024-12-09T20:53:11,127 INFO [Time-limited test {}] server.Server(415): Started @12598ms 2024-12-09T20:53:11,128 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T20:53:12,102 WARN [Thread-567 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data1/current/BP-216836135-172.17.0.2-1733777590115/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:12,103 WARN [Thread-568 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data2/current/BP-216836135-172.17.0.2-1733777590115/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:12,120 WARN [Thread-508 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T20:53:12,123 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3916c36c84ef837f with lease ID 0x59ea5cb820677e3a: Processing first storage report for DS-b21b2526-bd5c-4882-8e9b-8b65c322608a from datanode DatanodeRegistration(127.0.0.1:35583, datanodeUuid=c09e6b75-01c4-4ae2-aada-2124c1251ba2, infoPort=45839, infoSecurePort=0, ipcPort=33423, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115) 2024-12-09T20:53:12,123 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3916c36c84ef837f with lease ID 0x59ea5cb820677e3a: from storage DS-b21b2526-bd5c-4882-8e9b-8b65c322608a node DatanodeRegistration(127.0.0.1:35583, datanodeUuid=c09e6b75-01c4-4ae2-aada-2124c1251ba2, infoPort=45839, infoSecurePort=0, ipcPort=33423, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:12,123 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3916c36c84ef837f with lease ID 0x59ea5cb820677e3a: Processing first storage report for DS-82f82e2f-9b85-4508-b519-6888e581396e from datanode DatanodeRegistration(127.0.0.1:35583, datanodeUuid=c09e6b75-01c4-4ae2-aada-2124c1251ba2, infoPort=45839, infoSecurePort=0, ipcPort=33423, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115) 2024-12-09T20:53:12,123 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3916c36c84ef837f with lease ID 0x59ea5cb820677e3a: from storage DS-82f82e2f-9b85-4508-b519-6888e581396e node DatanodeRegistration(127.0.0.1:35583, datanodeUuid=c09e6b75-01c4-4ae2-aada-2124c1251ba2, infoPort=45839, infoSecurePort=0, ipcPort=33423, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:12,250 WARN [Thread-579 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data3/current/BP-216836135-172.17.0.2-1733777590115/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:12,250 WARN [Thread-580 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data4/current/BP-216836135-172.17.0.2-1733777590115/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:12,267 WARN [Thread-531 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T20:53:12,269 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x66731a30e48ba114 with lease ID 0x59ea5cb820677e3b: Processing first storage report for DS-0db0470e-d511-4888-b620-b9f86de41669 from datanode DatanodeRegistration(127.0.0.1:45413, datanodeUuid=61260aaa-ea76-4f6e-9008-27f31da5f457, infoPort=33195, infoSecurePort=0, ipcPort=36673, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115) 2024-12-09T20:53:12,270 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x66731a30e48ba114 with lease ID 0x59ea5cb820677e3b: from storage DS-0db0470e-d511-4888-b620-b9f86de41669 node DatanodeRegistration(127.0.0.1:45413, datanodeUuid=61260aaa-ea76-4f6e-9008-27f31da5f457, infoPort=33195, infoSecurePort=0, ipcPort=36673, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:12,270 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x66731a30e48ba114 with lease ID 0x59ea5cb820677e3b: Processing first storage report for DS-a88d4609-07dc-47f1-9279-18efccf32b4a from datanode DatanodeRegistration(127.0.0.1:45413, datanodeUuid=61260aaa-ea76-4f6e-9008-27f31da5f457, infoPort=33195, infoSecurePort=0, ipcPort=36673, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115) 2024-12-09T20:53:12,270 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x66731a30e48ba114 with lease ID 0x59ea5cb820677e3b: from storage DS-a88d4609-07dc-47f1-9279-18efccf32b4a node DatanodeRegistration(127.0.0.1:45413, datanodeUuid=61260aaa-ea76-4f6e-9008-27f31da5f457, infoPort=33195, infoSecurePort=0, ipcPort=36673, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:12,356 WARN [Thread-590 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data5/current/BP-216836135-172.17.0.2-1733777590115/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:12,356 WARN [Thread-591 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data6/current/BP-216836135-172.17.0.2-1733777590115/current, will proceed with Du for space computation calculation, 2024-12-09T20:53:12,377 WARN [Thread-553 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T20:53:12,379 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7148a4319f6fe134 with lease ID 0x59ea5cb820677e3c: Processing first storage report for DS-c19c66c7-5cc1-4e19-9138-471f272f12ef from datanode DatanodeRegistration(127.0.0.1:34569, datanodeUuid=6ca5fd4c-c694-47f2-8311-4d5ac60e4c1b, infoPort=43849, infoSecurePort=0, ipcPort=38175, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115) 2024-12-09T20:53:12,379 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7148a4319f6fe134 with lease ID 0x59ea5cb820677e3c: from storage DS-c19c66c7-5cc1-4e19-9138-471f272f12ef node DatanodeRegistration(127.0.0.1:34569, datanodeUuid=6ca5fd4c-c694-47f2-8311-4d5ac60e4c1b, infoPort=43849, infoSecurePort=0, ipcPort=38175, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T20:53:12,379 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7148a4319f6fe134 with lease ID 0x59ea5cb820677e3c: Processing first storage report for DS-920f7aff-4757-4815-94ad-e7fa406fedf3 from datanode DatanodeRegistration(127.0.0.1:34569, datanodeUuid=6ca5fd4c-c694-47f2-8311-4d5ac60e4c1b, infoPort=43849, infoSecurePort=0, ipcPort=38175, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115) 2024-12-09T20:53:12,379 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7148a4319f6fe134 with lease ID 0x59ea5cb820677e3c: from storage DS-920f7aff-4757-4815-94ad-e7fa406fedf3 node DatanodeRegistration(127.0.0.1:34569, datanodeUuid=6ca5fd4c-c694-47f2-8311-4d5ac60e4c1b, infoPort=43849, infoSecurePort=0, ipcPort=38175, storageInfo=lv=-57;cid=testClusterID;nsid=711125119;c=1733777590115), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-09T20:53:12,468 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4 2024-12-09T20:53:12,472 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/zookeeper_0, clientPort=62017, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-09T20:53:12,473 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=62017 2024-12-09T20:53:12,473 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,475 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,487 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741825_1001 (size=7) 2024-12-09T20:53:12,488 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741825_1001 (size=7) 2024-12-09T20:53:12,488 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741825_1001 (size=7) 2024-12-09T20:53:12,490 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9 with version=8 2024-12-09T20:53:12,490 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:33595/user/jenkins/test-data/36ae464c-dbcf-95ca-f28d-f12395e59147/hbase-staging 2024-12-09T20:53:12,492 INFO [Time-limited test {}] client.ConnectionUtils(128): master/00c23a7251f8:0 server-side Connection retries=45 2024-12-09T20:53:12,492 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,492 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,492 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T20:53:12,492 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,492 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T20:53:12,492 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-09T20:53:12,493 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T20:53:12,493 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:38673 2024-12-09T20:53:12,494 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:38673 connecting to ZooKeeper ensemble=127.0.0.1:62017 2024-12-09T20:53:12,542 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:386730x0, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T20:53:12,542 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:38673-0x1000c5a975e0000 connected 2024-12-09T20:53:12,620 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,626 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,630 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:12,631 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9, hbase.cluster.distributed=false 2024-12-09T20:53:12,633 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T20:53:12,634 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38673 2024-12-09T20:53:12,634 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38673 2024-12-09T20:53:12,634 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38673 2024-12-09T20:53:12,635 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38673 2024-12-09T20:53:12,635 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38673 2024-12-09T20:53:12,647 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c23a7251f8:0 server-side Connection retries=45 2024-12-09T20:53:12,648 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,648 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,648 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T20:53:12,648 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,648 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T20:53:12,648 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T20:53:12,648 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T20:53:12,649 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36251 2024-12-09T20:53:12,650 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36251 connecting to ZooKeeper ensemble=127.0.0.1:62017 2024-12-09T20:53:12,651 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,653 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,661 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:362510x0, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T20:53:12,662 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:12,662 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36251-0x1000c5a975e0001 connected 2024-12-09T20:53:12,662 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T20:53:12,662 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T20:53:12,663 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T20:53:12,664 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T20:53:12,665 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36251 2024-12-09T20:53:12,665 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36251 2024-12-09T20:53:12,666 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36251 2024-12-09T20:53:12,666 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36251 2024-12-09T20:53:12,666 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36251 2024-12-09T20:53:12,682 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c23a7251f8:0 server-side Connection retries=45 2024-12-09T20:53:12,682 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,682 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,683 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T20:53:12,683 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,683 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T20:53:12,683 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T20:53:12,683 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T20:53:12,683 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36855 2024-12-09T20:53:12,685 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36855 connecting to ZooKeeper ensemble=127.0.0.1:62017 2024-12-09T20:53:12,685 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,687 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,693 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:368550x0, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T20:53:12,693 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:12,693 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36855-0x1000c5a975e0002 connected 2024-12-09T20:53:12,693 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T20:53:12,694 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T20:53:12,694 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T20:53:12,695 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T20:53:12,696 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36855 2024-12-09T20:53:12,697 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36855 2024-12-09T20:53:12,697 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36855 2024-12-09T20:53:12,697 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36855 2024-12-09T20:53:12,697 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36855 2024-12-09T20:53:12,711 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c23a7251f8:0 server-side Connection retries=45 2024-12-09T20:53:12,711 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,711 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,711 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T20:53:12,711 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T20:53:12,711 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T20:53:12,712 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T20:53:12,712 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T20:53:12,712 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:43799 2024-12-09T20:53:12,714 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:43799 connecting to ZooKeeper ensemble=127.0.0.1:62017 2024-12-09T20:53:12,714 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,715 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,724 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:437990x0, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T20:53:12,725 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:437990x0, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:12,725 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:43799-0x1000c5a975e0003 connected 2024-12-09T20:53:12,725 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T20:53:12,726 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T20:53:12,726 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T20:53:12,728 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T20:53:12,728 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43799 2024-12-09T20:53:12,729 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43799 2024-12-09T20:53:12,729 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43799 2024-12-09T20:53:12,729 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43799 2024-12-09T20:53:12,729 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43799 2024-12-09T20:53:12,742 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;00c23a7251f8:38673 2024-12-09T20:53:12,742 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/00c23a7251f8,38673,1733777592492 2024-12-09T20:53:12,755 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:12,755 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:12,755 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:12,755 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:12,756 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/00c23a7251f8,38673,1733777592492 2024-12-09T20:53:12,766 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T20:53:12,766 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T20:53:12,766 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T20:53:12,766 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,767 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,767 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,767 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,768 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-09T20:53:12,769 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/00c23a7251f8,38673,1733777592492 from backup master directory 2024-12-09T20:53:12,777 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:12,777 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:12,777 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/00c23a7251f8,38673,1733777592492 2024-12-09T20:53:12,777 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:12,777 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T20:53:12,777 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T20:53:12,777 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=00c23a7251f8,38673,1733777592492 2024-12-09T20:53:12,784 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/hbase.id] with ID: 16a9bb92-b289-4876-9e7e-bda9d17902fd 2024-12-09T20:53:12,784 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/.tmp/hbase.id 2024-12-09T20:53:12,792 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741826_1002 (size=42) 2024-12-09T20:53:12,792 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741826_1002 (size=42) 2024-12-09T20:53:12,793 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741826_1002 (size=42) 2024-12-09T20:53:12,793 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/.tmp/hbase.id]:[hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/hbase.id] 2024-12-09T20:53:12,808 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T20:53:12,808 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-09T20:53:12,810 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-12-09T20:53:12,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741827_1003 (size=196) 2024-12-09T20:53:12,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741827_1003 (size=196) 2024-12-09T20:53:12,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741827_1003 (size=196) 2024-12-09T20:53:12,832 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-09T20:53:12,833 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-09T20:53:12,833 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T20:53:12,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741828_1004 (size=1189) 2024-12-09T20:53:12,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741828_1004 (size=1189) 2024-12-09T20:53:12,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741828_1004 (size=1189) 2024-12-09T20:53:12,845 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store 2024-12-09T20:53:12,852 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741829_1005 (size=34) 2024-12-09T20:53:12,853 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741829_1005 (size=34) 2024-12-09T20:53:12,853 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741829_1005 (size=34) 2024-12-09T20:53:12,854 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:12,854 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-09T20:53:12,854 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:12,854 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:12,854 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-09T20:53:12,854 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:12,854 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:12,854 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733777592854Disabling compacts and flushes for region at 1733777592854Disabling writes for close at 1733777592854Writing region close event to WAL at 1733777592854Closed at 1733777592854 2024-12-09T20:53:12,855 WARN [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/.initializing 2024-12-09T20:53:12,855 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/WALs/00c23a7251f8,38673,1733777592492 2024-12-09T20:53:12,858 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C38673%2C1733777592492, suffix=, logDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/WALs/00c23a7251f8,38673,1733777592492, archiveDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/oldWALs, maxLogs=10 2024-12-09T20:53:12,859 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c23a7251f8%2C38673%2C1733777592492.1733777592858 2024-12-09T20:53:12,869 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/WALs/00c23a7251f8,38673,1733777592492/00c23a7251f8%2C38673%2C1733777592492.1733777592858 2024-12-09T20:53:12,871 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43849:43849),(127.0.0.1/127.0.0.1:33195:33195),(127.0.0.1/127.0.0.1:45839:45839)] 2024-12-09T20:53:12,872 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-09T20:53:12,872 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:12,872 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,872 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,874 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,876 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-09T20:53:12,876 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:12,877 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:12,877 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,879 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-09T20:53:12,880 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:12,880 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T20:53:12,880 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,883 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-09T20:53:12,884 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:12,884 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T20:53:12,885 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,886 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-09T20:53:12,886 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:12,887 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T20:53:12,887 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,888 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,888 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,890 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,890 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,890 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T20:53:12,892 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T20:53:12,895 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T20:53:12,896 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61352315, jitterRate=-0.08577926456928253}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T20:53:12,897 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733777592872Initializing all the Stores at 1733777592874 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777592874Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777592874Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777592874Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777592874Cleaning up temporary data from old regions at 1733777592890 (+16 ms)Region opened successfully at 1733777592897 (+7 ms) 2024-12-09T20:53:12,897 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-09T20:53:12,902 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@47ca7930, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c23a7251f8/172.17.0.2:0 2024-12-09T20:53:12,903 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-09T20:53:12,903 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-09T20:53:12,903 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-09T20:53:12,904 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-09T20:53:12,904 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-09T20:53:12,905 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-09T20:53:12,905 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-09T20:53:12,908 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-09T20:53:12,909 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-09T20:53:12,913 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-09T20:53:12,914 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-09T20:53:12,915 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-09T20:53:12,919 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-09T20:53:12,924 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-09T20:53:12,924 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-09T20:53:12,925 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-09T20:53:12,926 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-09T20:53:12,935 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-09T20:53:12,936 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-09T20:53:12,945 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-09T20:53:12,948 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-09T20:53:12,956 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-09T20:53:12,959 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-09T20:53:12,960 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-09T20:53:12,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:12,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:12,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:12,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:12,967 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,967 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=00c23a7251f8,38673,1733777592492, sessionid=0x1000c5a975e0000, setting cluster-up flag (Was=false) 2024-12-09T20:53:12,987 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,987 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,987 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:12,987 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,019 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-09T20:53:13,023 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=00c23a7251f8,38673,1733777592492 2024-12-09T20:53:13,040 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,040 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,040 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,041 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,293 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-09T20:53:13,299 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=00c23a7251f8,38673,1733777592492 2024-12-09T20:53:13,301 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-09T20:53:13,304 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-09T20:53:13,304 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-09T20:53:13,304 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-09T20:53:13,305 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 00c23a7251f8,38673,1733777592492 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-09T20:53:13,307 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/00c23a7251f8:0, corePoolSize=5, maxPoolSize=5 2024-12-09T20:53:13,307 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/00c23a7251f8:0, corePoolSize=5, maxPoolSize=5 2024-12-09T20:53:13,307 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/00c23a7251f8:0, corePoolSize=5, maxPoolSize=5 2024-12-09T20:53:13,307 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/00c23a7251f8:0, corePoolSize=5, maxPoolSize=5 2024-12-09T20:53:13,307 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/00c23a7251f8:0, corePoolSize=10, maxPoolSize=10 2024-12-09T20:53:13,307 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,307 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/00c23a7251f8:0, corePoolSize=2, maxPoolSize=2 2024-12-09T20:53:13,307 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,308 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733777623308 2024-12-09T20:53:13,308 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-09T20:53:13,308 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-09T20:53:13,308 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-09T20:53:13,308 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-09T20:53:13,309 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-09T20:53:13,309 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-09T20:53:13,309 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,309 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-09T20:53:13,309 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-09T20:53:13,309 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-09T20:53:13,310 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T20:53:13,310 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-09T20:53:13,310 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-09T20:53:13,310 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-09T20:53:13,310 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.large.0-1733777593310,5,FailOnTimeoutGroup] 2024-12-09T20:53:13,310 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.small.0-1733777593310,5,FailOnTimeoutGroup] 2024-12-09T20:53:13,310 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,310 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-09T20:53:13,310 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,310 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,312 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:13,312 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-09T20:53:13,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741831_1007 (size=1321) 2024-12-09T20:53:13,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741831_1007 (size=1321) 2024-12-09T20:53:13,322 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741831_1007 (size=1321) 2024-12-09T20:53:13,322 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-09T20:53:13,323 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9 2024-12-09T20:53:13,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741832_1008 (size=32) 2024-12-09T20:53:13,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741832_1008 (size=32) 2024-12-09T20:53:13,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741832_1008 (size=32) 2024-12-09T20:53:13,332 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(746): ClusterId : 16a9bb92-b289-4876-9e7e-bda9d17902fd 2024-12-09T20:53:13,332 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(746): ClusterId : 16a9bb92-b289-4876-9e7e-bda9d17902fd 2024-12-09T20:53:13,332 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T20:53:13,332 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(746): ClusterId : 16a9bb92-b289-4876-9e7e-bda9d17902fd 2024-12-09T20:53:13,332 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T20:53:13,332 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T20:53:13,333 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:13,334 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-09T20:53:13,336 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-09T20:53:13,337 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:13,337 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:13,337 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-09T20:53:13,339 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-09T20:53:13,339 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:13,340 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:13,340 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-09T20:53:13,342 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-09T20:53:13,342 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:13,342 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:13,342 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-09T20:53:13,344 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-09T20:53:13,344 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:13,345 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:13,345 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-09T20:53:13,346 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740 2024-12-09T20:53:13,346 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740 2024-12-09T20:53:13,348 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-09T20:53:13,348 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-09T20:53:13,349 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T20:53:13,350 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-09T20:53:13,352 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T20:53:13,353 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67775212, jitterRate=0.009929358959197998}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T20:53:13,354 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733777593333Initializing all the Stores at 1733777593334 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777593334Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777593334Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777593334Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777593334Cleaning up temporary data from old regions at 1733777593348 (+14 ms)Region opened successfully at 1733777593354 (+6 ms) 2024-12-09T20:53:13,354 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-09T20:53:13,354 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-09T20:53:13,354 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-09T20:53:13,354 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-09T20:53:13,354 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-09T20:53:13,354 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-09T20:53:13,354 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733777593354Disabling compacts and flushes for region at 1733777593354Disabling writes for close at 1733777593354Writing region close event to WAL at 1733777593354Closed at 1733777593354 2024-12-09T20:53:13,356 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T20:53:13,356 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T20:53:13,356 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T20:53:13,356 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-09T20:53:13,356 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T20:53:13,356 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T20:53:13,356 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-09T20:53:13,357 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T20:53:13,357 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T20:53:13,358 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-09T20:53:13,359 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-09T20:53:13,377 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T20:53:13,378 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T20:53:13,378 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T20:53:13,378 DEBUG [RS:0;00c23a7251f8:36251 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@394e39d3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c23a7251f8/172.17.0.2:0 2024-12-09T20:53:13,378 DEBUG [RS:2;00c23a7251f8:43799 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@720246ac, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c23a7251f8/172.17.0.2:0 2024-12-09T20:53:13,378 DEBUG [RS:1;00c23a7251f8:36855 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@290d45c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c23a7251f8/172.17.0.2:0 2024-12-09T20:53:13,388 DEBUG [RS:2;00c23a7251f8:43799 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;00c23a7251f8:43799 2024-12-09T20:53:13,388 DEBUG [RS:1;00c23a7251f8:36855 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;00c23a7251f8:36855 2024-12-09T20:53:13,388 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T20:53:13,388 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T20:53:13,388 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T20:53:13,388 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T20:53:13,388 DEBUG [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T20:53:13,388 DEBUG [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T20:53:13,389 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c23a7251f8,38673,1733777592492 with port=43799, startcode=1733777592711 2024-12-09T20:53:13,389 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c23a7251f8,38673,1733777592492 with port=36855, startcode=1733777592682 2024-12-09T20:53:13,389 DEBUG [RS:2;00c23a7251f8:43799 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T20:53:13,389 DEBUG [RS:1;00c23a7251f8:36855 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T20:53:13,392 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33945, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T20:53:13,393 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38673 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c23a7251f8,43799,1733777592711 2024-12-09T20:53:13,393 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38673 {}] master.ServerManager(517): Registering regionserver=00c23a7251f8,43799,1733777592711 2024-12-09T20:53:13,393 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:57661, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T20:53:13,393 DEBUG [RS:0;00c23a7251f8:36251 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;00c23a7251f8:36251 2024-12-09T20:53:13,393 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T20:53:13,393 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T20:53:13,394 DEBUG [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T20:53:13,394 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c23a7251f8,38673,1733777592492 with port=36251, startcode=1733777592647 2024-12-09T20:53:13,395 DEBUG [RS:0;00c23a7251f8:36251 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T20:53:13,395 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38673 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c23a7251f8,36855,1733777592682 2024-12-09T20:53:13,395 DEBUG [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9 2024-12-09T20:53:13,395 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38673 {}] master.ServerManager(517): Registering regionserver=00c23a7251f8,36855,1733777592682 2024-12-09T20:53:13,395 DEBUG [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33191 2024-12-09T20:53:13,395 DEBUG [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T20:53:13,396 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37751, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T20:53:13,398 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38673 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c23a7251f8,36251,1733777592647 2024-12-09T20:53:13,398 DEBUG [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9 2024-12-09T20:53:13,398 DEBUG [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33191 2024-12-09T20:53:13,398 DEBUG [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T20:53:13,398 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38673 {}] master.ServerManager(517): Registering regionserver=00c23a7251f8,36251,1733777592647 2024-12-09T20:53:13,400 DEBUG [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9 2024-12-09T20:53:13,400 DEBUG [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33191 2024-12-09T20:53:13,400 DEBUG [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T20:53:13,408 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T20:53:13,482 DEBUG [RS:2;00c23a7251f8:43799 {}] zookeeper.ZKUtil(111): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c23a7251f8,43799,1733777592711 2024-12-09T20:53:13,482 WARN [RS:2;00c23a7251f8:43799 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T20:53:13,482 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c23a7251f8,43799,1733777592711] 2024-12-09T20:53:13,482 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c23a7251f8,36855,1733777592682] 2024-12-09T20:53:13,482 INFO [RS:2;00c23a7251f8:43799 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T20:53:13,482 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c23a7251f8,36251,1733777592647] 2024-12-09T20:53:13,482 DEBUG [RS:0;00c23a7251f8:36251 {}] zookeeper.ZKUtil(111): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c23a7251f8,36251,1733777592647 2024-12-09T20:53:13,482 DEBUG [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,43799,1733777592711 2024-12-09T20:53:13,482 WARN [RS:0;00c23a7251f8:36251 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T20:53:13,482 DEBUG [RS:1;00c23a7251f8:36855 {}] zookeeper.ZKUtil(111): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c23a7251f8,36855,1733777592682 2024-12-09T20:53:13,482 INFO [RS:0;00c23a7251f8:36251 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T20:53:13,482 WARN [RS:1;00c23a7251f8:36855 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T20:53:13,482 DEBUG [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,36251,1733777592647 2024-12-09T20:53:13,482 INFO [RS:1;00c23a7251f8:36855 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T20:53:13,482 DEBUG [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,36855,1733777592682 2024-12-09T20:53:13,486 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T20:53:13,486 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T20:53:13,486 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T20:53:13,489 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T20:53:13,489 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T20:53:13,491 INFO [RS:2;00c23a7251f8:43799 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T20:53:13,491 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,492 INFO [RS:0;00c23a7251f8:36251 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T20:53:13,492 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T20:53:13,492 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T20:53:13,492 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,493 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T20:53:13,493 INFO [RS:1;00c23a7251f8:36855 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T20:53:13,493 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,493 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T20:53:13,493 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T20:53:13,493 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,493 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,493 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T20:53:13,493 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,493 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,493 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c23a7251f8:0, corePoolSize=2, maxPoolSize=2 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c23a7251f8:0, corePoolSize=2, maxPoolSize=2 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:13,494 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c23a7251f8:0, corePoolSize=2, maxPoolSize=2 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:2;00c23a7251f8:43799 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,494 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,495 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:13,495 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,495 DEBUG [RS:0;00c23a7251f8:36251 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:13,495 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,495 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,495 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c23a7251f8:0, corePoolSize=1, maxPoolSize=1 2024-12-09T20:53:13,495 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:13,495 DEBUG [RS:1;00c23a7251f8:36855 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0, corePoolSize=3, maxPoolSize=3 2024-12-09T20:53:13,497 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,497 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,497 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,497 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,497 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,497 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,43799,1733777592711-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T20:53:13,500 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,500 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,500 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,500 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,500 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,500 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36251,1733777592647-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T20:53:13,500 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,501 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,501 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,501 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,501 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,501 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36855,1733777592682-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T20:53:13,510 WARN [00c23a7251f8:38673 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-09T20:53:13,513 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T20:53:13,513 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T20:53:13,513 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,43799,1733777592711-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,513 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,513 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36251,1733777592647-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,513 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.Replication(171): 00c23a7251f8,43799,1733777592711 started 2024-12-09T20:53:13,513 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,513 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.Replication(171): 00c23a7251f8,36251,1733777592647 started 2024-12-09T20:53:13,519 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T20:53:13,520 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,36855,1733777592682-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,520 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,520 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.Replication(171): 00c23a7251f8,36855,1733777592682 started 2024-12-09T20:53:13,526 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,527 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1482): Serving as 00c23a7251f8,36251,1733777592647, RpcServer on 00c23a7251f8/172.17.0.2:36251, sessionid=0x1000c5a975e0001 2024-12-09T20:53:13,527 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T20:53:13,527 DEBUG [RS:0;00c23a7251f8:36251 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c23a7251f8,36251,1733777592647 2024-12-09T20:53:13,527 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,36251,1733777592647' 2024-12-09T20:53:13,527 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,527 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T20:53:13,527 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(1482): Serving as 00c23a7251f8,43799,1733777592711, RpcServer on 00c23a7251f8/172.17.0.2:43799, sessionid=0x1000c5a975e0003 2024-12-09T20:53:13,527 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T20:53:13,527 DEBUG [RS:2;00c23a7251f8:43799 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c23a7251f8,43799,1733777592711 2024-12-09T20:53:13,527 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,43799,1733777592711' 2024-12-09T20:53:13,527 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T20:53:13,528 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T20:53:13,528 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T20:53:13,528 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T20:53:13,528 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T20:53:13,528 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T20:53:13,528 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T20:53:13,528 DEBUG [RS:2;00c23a7251f8:43799 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c23a7251f8,43799,1733777592711 2024-12-09T20:53:13,528 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,43799,1733777592711' 2024-12-09T20:53:13,528 DEBUG [RS:0;00c23a7251f8:36251 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c23a7251f8,36251,1733777592647 2024-12-09T20:53:13,528 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T20:53:13,528 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,36251,1733777592647' 2024-12-09T20:53:13,528 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T20:53:13,529 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T20:53:13,529 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T20:53:13,529 DEBUG [RS:2;00c23a7251f8:43799 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T20:53:13,529 INFO [RS:2;00c23a7251f8:43799 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T20:53:13,529 INFO [RS:2;00c23a7251f8:43799 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T20:53:13,529 DEBUG [RS:0;00c23a7251f8:36251 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T20:53:13,529 INFO [RS:0;00c23a7251f8:36251 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T20:53:13,529 INFO [RS:0;00c23a7251f8:36251 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T20:53:13,538 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:13,539 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1482): Serving as 00c23a7251f8,36855,1733777592682, RpcServer on 00c23a7251f8/172.17.0.2:36855, sessionid=0x1000c5a975e0002 2024-12-09T20:53:13,539 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T20:53:13,539 DEBUG [RS:1;00c23a7251f8:36855 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c23a7251f8,36855,1733777592682 2024-12-09T20:53:13,539 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,36855,1733777592682' 2024-12-09T20:53:13,539 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T20:53:13,540 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T20:53:13,540 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T20:53:13,540 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T20:53:13,540 DEBUG [RS:1;00c23a7251f8:36855 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c23a7251f8,36855,1733777592682 2024-12-09T20:53:13,540 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c23a7251f8,36855,1733777592682' 2024-12-09T20:53:13,540 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T20:53:13,541 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T20:53:13,541 DEBUG [RS:1;00c23a7251f8:36855 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T20:53:13,541 INFO [RS:1;00c23a7251f8:36855 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T20:53:13,541 INFO [RS:1;00c23a7251f8:36855 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T20:53:13,634 INFO [RS:2;00c23a7251f8:43799 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C43799%2C1733777592711, suffix=, logDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,43799,1733777592711, archiveDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/oldWALs, maxLogs=32 2024-12-09T20:53:13,634 INFO [RS:0;00c23a7251f8:36251 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C36251%2C1733777592647, suffix=, logDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,36251,1733777592647, archiveDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/oldWALs, maxLogs=32 2024-12-09T20:53:13,638 INFO [RS:2;00c23a7251f8:43799 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c23a7251f8%2C43799%2C1733777592711.1733777593638 2024-12-09T20:53:13,638 INFO [RS:0;00c23a7251f8:36251 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c23a7251f8%2C36251%2C1733777592647.1733777593638 2024-12-09T20:53:13,643 INFO [RS:1;00c23a7251f8:36855 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C36855%2C1733777592682, suffix=, logDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,36855,1733777592682, archiveDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/oldWALs, maxLogs=32 2024-12-09T20:53:13,645 INFO [RS:1;00c23a7251f8:36855 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c23a7251f8%2C36855%2C1733777592682.1733777593645 2024-12-09T20:53:13,648 INFO [RS:2;00c23a7251f8:43799 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,43799,1733777592711/00c23a7251f8%2C43799%2C1733777592711.1733777593638 2024-12-09T20:53:13,649 INFO [RS:0;00c23a7251f8:36251 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,36251,1733777592647/00c23a7251f8%2C36251%2C1733777592647.1733777593638 2024-12-09T20:53:13,650 DEBUG [RS:2;00c23a7251f8:43799 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43849:43849),(127.0.0.1/127.0.0.1:33195:33195),(127.0.0.1/127.0.0.1:45839:45839)] 2024-12-09T20:53:13,650 DEBUG [RS:0;00c23a7251f8:36251 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43849:43849),(127.0.0.1/127.0.0.1:33195:33195),(127.0.0.1/127.0.0.1:45839:45839)] 2024-12-09T20:53:13,654 INFO [RS:1;00c23a7251f8:36855 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,36855,1733777592682/00c23a7251f8%2C36855%2C1733777592682.1733777593645 2024-12-09T20:53:13,655 DEBUG [RS:1;00c23a7251f8:36855 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:33195:33195),(127.0.0.1/127.0.0.1:45839:45839),(127.0.0.1/127.0.0.1:43849:43849)] 2024-12-09T20:53:13,760 DEBUG [00c23a7251f8:38673 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-09T20:53:13,761 DEBUG [00c23a7251f8:38673 {}] balancer.BalancerClusterState(204): Hosts are {00c23a7251f8=0} racks are {/default-rack=0} 2024-12-09T20:53:13,765 DEBUG [00c23a7251f8:38673 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-09T20:53:13,765 DEBUG [00c23a7251f8:38673 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-09T20:53:13,765 DEBUG [00c23a7251f8:38673 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-09T20:53:13,765 DEBUG [00c23a7251f8:38673 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-09T20:53:13,765 DEBUG [00c23a7251f8:38673 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-09T20:53:13,765 DEBUG [00c23a7251f8:38673 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-09T20:53:13,765 INFO [00c23a7251f8:38673 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-09T20:53:13,765 INFO [00c23a7251f8:38673 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-09T20:53:13,765 INFO [00c23a7251f8:38673 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-09T20:53:13,765 DEBUG [00c23a7251f8:38673 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-09T20:53:13,766 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=00c23a7251f8,36855,1733777592682 2024-12-09T20:53:13,769 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 00c23a7251f8,36855,1733777592682, state=OPENING 2024-12-09T20:53:13,798 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-09T20:53:13,809 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,809 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,809 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,809 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:13,810 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-09T20:53:13,811 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:13,811 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:13,811 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:13,811 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:13,811 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=00c23a7251f8,36855,1733777592682}] 2024-12-09T20:53:13,969 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-09T20:53:13,972 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36211, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-09T20:53:13,977 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-09T20:53:13,978 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T20:53:13,980 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c23a7251f8%2C36855%2C1733777592682.meta, suffix=.meta, logDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,36855,1733777592682, archiveDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/oldWALs, maxLogs=32 2024-12-09T20:53:13,981 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c23a7251f8%2C36855%2C1733777592682.meta.1733777593981.meta 2024-12-09T20:53:13,988 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/WALs/00c23a7251f8,36855,1733777592682/00c23a7251f8%2C36855%2C1733777592682.meta.1733777593981.meta 2024-12-09T20:53:13,991 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45839:45839),(127.0.0.1/127.0.0.1:43849:43849),(127.0.0.1/127.0.0.1:33195:33195)] 2024-12-09T20:53:13,992 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-09T20:53:13,992 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-09T20:53:13,992 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-09T20:53:13,992 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-09T20:53:13,993 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-09T20:53:13,993 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:13,993 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-09T20:53:13,993 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-09T20:53:13,995 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-09T20:53:13,996 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-09T20:53:13,996 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:13,996 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:13,997 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-09T20:53:13,998 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-09T20:53:13,998 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:13,998 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:13,999 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-09T20:53:14,000 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-09T20:53:14,000 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:14,000 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:14,000 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-09T20:53:14,002 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-09T20:53:14,002 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:14,002 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T20:53:14,003 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-09T20:53:14,004 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740 2024-12-09T20:53:14,005 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740 2024-12-09T20:53:14,007 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-09T20:53:14,007 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-09T20:53:14,008 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T20:53:14,010 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-09T20:53:14,011 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=75158957, jitterRate=0.11995573341846466}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T20:53:14,011 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-09T20:53:14,013 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733777593993Writing region info on filesystem at 1733777593993Initializing all the Stores at 1733777593994 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777593994Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777593995 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777593995Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733777593995Cleaning up temporary data from old regions at 1733777594007 (+12 ms)Running coprocessor post-open hooks at 1733777594011 (+4 ms)Region opened successfully at 1733777594012 (+1 ms) 2024-12-09T20:53:14,014 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733777593968 2024-12-09T20:53:14,018 DEBUG [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-09T20:53:14,018 INFO [RS_OPEN_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-09T20:53:14,019 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=00c23a7251f8,36855,1733777592682 2024-12-09T20:53:14,021 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 00c23a7251f8,36855,1733777592682, state=OPEN 2024-12-09T20:53:14,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T20:53:14,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T20:53:14,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T20:53:14,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T20:53:14,104 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=00c23a7251f8,36855,1733777592682 2024-12-09T20:53:14,104 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:14,104 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:14,104 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:14,104 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T20:53:14,113 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-09T20:53:14,113 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=00c23a7251f8,36855,1733777592682 in 293 msec 2024-12-09T20:53:14,117 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-09T20:53:14,117 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 757 msec 2024-12-09T20:53:14,118 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T20:53:14,119 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-09T20:53:14,120 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-09T20:53:14,120 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=00c23a7251f8,36855,1733777592682, seqNum=-1] 2024-12-09T20:53:14,121 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T20:53:14,122 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48679, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T20:53:14,132 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 827 msec 2024-12-09T20:53:14,133 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733777594133, completionTime=-1 2024-12-09T20:53:14,133 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-09T20:53:14,133 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-09T20:53:14,135 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-09T20:53:14,135 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733777654135 2024-12-09T20:53:14,135 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733777714135 2024-12-09T20:53:14,136 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-12-09T20:53:14,136 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-09T20:53:14,136 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,38673,1733777592492-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:14,136 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,38673,1733777592492-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:14,136 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,38673,1733777592492-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:14,136 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-00c23a7251f8:38673, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:14,137 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:14,137 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:14,139 DEBUG [master/00c23a7251f8:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-09T20:53:14,142 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.365sec 2024-12-09T20:53:14,142 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-09T20:53:14,142 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-09T20:53:14,142 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-09T20:53:14,142 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-09T20:53:14,142 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-09T20:53:14,142 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,38673,1733777592492-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T20:53:14,142 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,38673,1733777592492-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-09T20:53:14,145 DEBUG [master/00c23a7251f8:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-09T20:53:14,145 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-09T20:53:14,145 INFO [master/00c23a7251f8:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c23a7251f8,38673,1733777592492-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T20:53:14,234 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@74e402b7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T20:53:14,234 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 00c23a7251f8,38673,-1 for getting cluster id 2024-12-09T20:53:14,235 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-09T20:53:14,237 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '16a9bb92-b289-4876-9e7e-bda9d17902fd' 2024-12-09T20:53:14,238 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-09T20:53:14,239 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "16a9bb92-b289-4876-9e7e-bda9d17902fd" 2024-12-09T20:53:14,240 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5b9431a7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T20:53:14,240 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [00c23a7251f8,38673,-1] 2024-12-09T20:53:14,240 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-09T20:53:14,241 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:14,244 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45200, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-09T20:53:14,246 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@77c67938, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T20:53:14,246 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-09T20:53:14,248 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=00c23a7251f8,36855,1733777592682, seqNum=-1] 2024-12-09T20:53:14,248 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T20:53:14,251 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54630, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T20:53:14,253 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=00c23a7251f8,38673,1733777592492 2024-12-09T20:53:14,253 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-09T20:53:14,254 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is 00c23a7251f8,38673,1733777592492 2024-12-09T20:53:14,254 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@5d3f14f4 2024-12-09T20:53:14,255 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-09T20:53:14,256 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45216, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-09T20:53:14,257 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-09T20:53:14,259 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-09T20:53:14,260 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-09T20:53:14,261 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:14,261 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-09T20:53:14,262 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:14,263 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-09T20:53:14,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741837_1013 (size=392) 2024-12-09T20:53:14,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741837_1013 (size=392) 2024-12-09T20:53:14,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741837_1013 (size=392) 2024-12-09T20:53:14,275 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 10382e18560e2242c4008357cc131575, NAME => 'TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9 2024-12-09T20:53:14,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741838_1014 (size=51) 2024-12-09T20:53:14,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741838_1014 (size=51) 2024-12-09T20:53:14,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741838_1014 (size=51) 2024-12-09T20:53:14,285 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:14,285 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 10382e18560e2242c4008357cc131575, disabling compactions & flushes 2024-12-09T20:53:14,285 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:14,285 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:14,285 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. after waiting 0 ms 2024-12-09T20:53:14,285 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:14,285 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:14,285 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 10382e18560e2242c4008357cc131575: Waiting for close lock at 1733777594285Disabling compacts and flushes for region at 1733777594285Disabling writes for close at 1733777594285Writing region close event to WAL at 1733777594285Closed at 1733777594285 2024-12-09T20:53:14,287 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-09T20:53:14,288 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733777594288"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733777594288"}]},"ts":"1733777594288"} 2024-12-09T20:53:14,291 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-09T20:53:14,293 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-09T20:53:14,293 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733777594293"}]},"ts":"1733777594293"} 2024-12-09T20:53:14,296 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-09T20:53:14,297 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {00c23a7251f8=0} racks are {/default-rack=0} 2024-12-09T20:53:14,298 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-09T20:53:14,298 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-09T20:53:14,298 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-09T20:53:14,298 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-09T20:53:14,298 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-09T20:53:14,298 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-09T20:53:14,298 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-09T20:53:14,298 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-09T20:53:14,298 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-09T20:53:14,298 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-09T20:53:14,298 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=10382e18560e2242c4008357cc131575, ASSIGN}] 2024-12-09T20:53:14,300 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=10382e18560e2242c4008357cc131575, ASSIGN 2024-12-09T20:53:14,302 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=10382e18560e2242c4008357cc131575, ASSIGN; state=OFFLINE, location=00c23a7251f8,36251,1733777592647; forceNewPlan=false, retain=false 2024-12-09T20:53:14,373 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:14,453 INFO [00c23a7251f8:38673 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-09T20:53:14,454 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=10382e18560e2242c4008357cc131575, regionState=OPENING, regionLocation=00c23a7251f8,36251,1733777592647 2024-12-09T20:53:14,463 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=10382e18560e2242c4008357cc131575, ASSIGN because future has completed 2024-12-09T20:53:14,463 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 10382e18560e2242c4008357cc131575, server=00c23a7251f8,36251,1733777592647}] 2024-12-09T20:53:14,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:14,618 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-09T20:53:14,619 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49251, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-09T20:53:14,624 INFO [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:14,624 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 10382e18560e2242c4008357cc131575, NAME => 'TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575.', STARTKEY => '', ENDKEY => ''} 2024-12-09T20:53:14,624 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,624 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T20:53:14,624 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,624 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,626 INFO [StoreOpener-10382e18560e2242c4008357cc131575-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,628 INFO [StoreOpener-10382e18560e2242c4008357cc131575-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 10382e18560e2242c4008357cc131575 columnFamilyName cf 2024-12-09T20:53:14,628 DEBUG [StoreOpener-10382e18560e2242c4008357cc131575-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T20:53:14,629 INFO [StoreOpener-10382e18560e2242c4008357cc131575-1 {}] regionserver.HStore(327): Store=10382e18560e2242c4008357cc131575/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T20:53:14,629 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,630 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,630 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,631 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,631 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,633 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,635 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T20:53:14,636 INFO [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 10382e18560e2242c4008357cc131575; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63988227, jitterRate=-0.04650111496448517}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-09T20:53:14,636 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 10382e18560e2242c4008357cc131575 2024-12-09T20:53:14,637 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 10382e18560e2242c4008357cc131575: Running coprocessor pre-open hook at 1733777594625Writing region info on filesystem at 1733777594625Initializing all the Stores at 1733777594626 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733777594626Cleaning up temporary data from old regions at 1733777594631 (+5 ms)Running coprocessor post-open hooks at 1733777594636 (+5 ms)Region opened successfully at 1733777594637 (+1 ms) 2024-12-09T20:53:14,638 INFO [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575., pid=6, masterSystemTime=1733777594617 2024-12-09T20:53:14,641 DEBUG [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:14,642 INFO [RS_OPEN_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:14,643 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=10382e18560e2242c4008357cc131575, regionState=OPEN, openSeqNum=2, regionLocation=00c23a7251f8,36251,1733777592647 2024-12-09T20:53:14,646 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 10382e18560e2242c4008357cc131575, server=00c23a7251f8,36251,1733777592647 because future has completed 2024-12-09T20:53:14,651 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-09T20:53:14,653 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 10382e18560e2242c4008357cc131575, server=00c23a7251f8,36251,1733777592647 in 185 msec 2024-12-09T20:53:14,655 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-09T20:53:14,655 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=10382e18560e2242c4008357cc131575, ASSIGN in 353 msec 2024-12-09T20:53:14,657 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-09T20:53:14,657 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733777594657"}]},"ts":"1733777594657"} 2024-12-09T20:53:14,660 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-09T20:53:14,662 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-09T20:53:14,665 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 405 msec 2024-12-09T20:53:14,892 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T20:53:14,893 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-09T20:53:14,893 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-09T20:53:14,893 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-09T20:53:14,897 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-09T20:53:14,897 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-09T20:53:14,897 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-09T20:53:14,900 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575., hostname=00c23a7251f8,36251,1733777592647, seqNum=2] 2024-12-09T20:53:14,900 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T20:53:14,902 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59216, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T20:53:14,906 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-09T20:53:14,906 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-09T20:53:14,906 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-09T20:53:14,908 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-09T20:53:14,909 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-09T20:53:14,909 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-09T20:53:14,910 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T20:53:14,910 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-09T20:53:14,910 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-09T20:53:14,910 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-09T20:53:14,912 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestHBaseWalOnEC 2024-12-09T20:53:14,912 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestHBaseWalOnEC Metrics about Tables on a single HBase RegionServer 2024-12-09T20:53:14,912 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-09T20:53:14,913 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-09T20:53:15,013 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T20:53:15,070 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36251 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-09T20:53:15,071 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:15,071 INFO [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 10382e18560e2242c4008357cc131575 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-09T20:53:15,094 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575/.tmp/cf/2415a3fc75f74c9c8bb4d926df842e44 is 36, key is row/cf:cq/1733777594903/Put/seqid=0 2024-12-09T20:53:15,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741839_1015 (size=4787) 2024-12-09T20:53:15,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741839_1015 (size=4787) 2024-12-09T20:53:15,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741839_1015 (size=4787) 2024-12-09T20:53:15,102 INFO [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575/.tmp/cf/2415a3fc75f74c9c8bb4d926df842e44 2024-12-09T20:53:15,112 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575/.tmp/cf/2415a3fc75f74c9c8bb4d926df842e44 as hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575/cf/2415a3fc75f74c9c8bb4d926df842e44 2024-12-09T20:53:15,120 INFO [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575/cf/2415a3fc75f74c9c8bb4d926df842e44, entries=1, sequenceid=5, filesize=4.7 K 2024-12-09T20:53:15,121 INFO [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 10382e18560e2242c4008357cc131575 in 50ms, sequenceid=5, compaction requested=false 2024-12-09T20:53:15,121 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 10382e18560e2242c4008357cc131575: 2024-12-09T20:53:15,121 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:15,121 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c23a7251f8:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-09T20:53:15,122 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-09T20:53:15,126 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-09T20:53:15,126 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 211 msec 2024-12-09T20:53:15,129 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 221 msec 2024-12-09T20:53:15,223 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38673 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T20:53:15,224 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-09T20:53:15,228 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-09T20:53:15,228 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-09T20:53:15,229 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:15,229 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,229 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,229 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-09T20:53:15,229 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-09T20:53:15,229 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1155378738, stopped=false 2024-12-09T20:53:15,229 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=00c23a7251f8,38673,1733777592492 2024-12-09T20:53:15,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:15,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:15,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:15,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T20:53:15,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:15,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:15,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:15,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:15,282 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-09T20:53:15,282 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-09T20:53:15,282 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:15,283 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:15,283 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,283 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:15,283 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c23a7251f8,36251,1733777592647' ***** 2024-12-09T20:53:15,283 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T20:53:15,283 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:15,283 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c23a7251f8,36855,1733777592682' ***** 2024-12-09T20:53:15,283 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T20:53:15,283 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T20:53:15,283 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T20:53:15,283 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T20:53:15,283 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c23a7251f8,43799,1733777592711' ***** 2024-12-09T20:53:15,283 INFO [RS:0;00c23a7251f8:36251 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T20:53:15,283 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T20:53:15,284 INFO [RS:0;00c23a7251f8:36251 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T20:53:15,284 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(3091): Received CLOSE for 10382e18560e2242c4008357cc131575 2024-12-09T20:53:15,284 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T20:53:15,284 INFO [RS:1;00c23a7251f8:36855 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T20:53:15,284 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T20:53:15,284 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T20:53:15,284 INFO [RS:1;00c23a7251f8:36855 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T20:53:15,284 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T20:53:15,284 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(959): stopping server 00c23a7251f8,36251,1733777592647 2024-12-09T20:53:15,284 INFO [RS:2;00c23a7251f8:43799 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T20:53:15,284 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(959): stopping server 00c23a7251f8,36855,1733777592682 2024-12-09T20:53:15,285 INFO [RS:0;00c23a7251f8:36251 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T20:53:15,285 INFO [RS:2;00c23a7251f8:43799 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T20:53:15,285 INFO [RS:1;00c23a7251f8:36855 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T20:53:15,285 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(959): stopping server 00c23a7251f8,43799,1733777592711 2024-12-09T20:53:15,285 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 10382e18560e2242c4008357cc131575, disabling compactions & flushes 2024-12-09T20:53:15,285 INFO [RS:2;00c23a7251f8:43799 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T20:53:15,285 INFO [RS:1;00c23a7251f8:36855 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;00c23a7251f8:36855. 2024-12-09T20:53:15,285 INFO [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:15,285 INFO [RS:2;00c23a7251f8:43799 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;00c23a7251f8:43799. 2024-12-09T20:53:15,285 DEBUG [RS:1;00c23a7251f8:36855 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:15,285 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:15,285 DEBUG [RS:1;00c23a7251f8:36855 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,285 DEBUG [RS:2;00c23a7251f8:43799 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:15,285 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. after waiting 0 ms 2024-12-09T20:53:15,285 DEBUG [RS:2;00c23a7251f8:43799 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,285 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:15,285 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T20:53:15,285 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T20:53:15,285 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(976): stopping server 00c23a7251f8,43799,1733777592711; all regions closed. 2024-12-09T20:53:15,285 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T20:53:15,285 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-09T20:53:15,285 INFO [RS:0;00c23a7251f8:36251 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;00c23a7251f8:36251. 2024-12-09T20:53:15,285 DEBUG [RS:0;00c23a7251f8:36251 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T20:53:15,285 DEBUG [RS:0;00c23a7251f8:36251 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,286 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-09T20:53:15,286 DEBUG [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-09T20:53:15,286 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-09T20:53:15,286 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,286 DEBUG [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1325): Online Regions={10382e18560e2242c4008357cc131575=TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575.} 2024-12-09T20:53:15,286 DEBUG [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-09T20:53:15,286 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-09T20:53:15,286 DEBUG [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1351): Waiting on 10382e18560e2242c4008357cc131575 2024-12-09T20:53:15,286 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,286 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-09T20:53:15,286 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-09T20:53:15,286 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,286 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-09T20:53:15,286 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-09T20:53:15,286 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,286 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,286 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-09T20:53:15,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741834_1010 (size=93) 2024-12-09T20:53:15,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741834_1010 (size=93) 2024-12-09T20:53:15,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741834_1010 (size=93) 2024-12-09T20:53:15,291 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/default/TestHBaseWalOnEC/10382e18560e2242c4008357cc131575/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-09T20:53:15,292 INFO [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:15,292 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 10382e18560e2242c4008357cc131575: Waiting for close lock at 1733777595285Running coprocessor pre-close hooks at 1733777595285Disabling compacts and flushes for region at 1733777595285Disabling writes for close at 1733777595285Writing region close event to WAL at 1733777595286 (+1 ms)Running coprocessor post-close hooks at 1733777595292 (+6 ms)Closed at 1733777595292 2024-12-09T20:53:15,292 DEBUG [RS_CLOSE_REGION-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575. 2024-12-09T20:53:15,294 DEBUG [RS:2;00c23a7251f8:43799 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/oldWALs 2024-12-09T20:53:15,294 INFO [RS:2;00c23a7251f8:43799 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 00c23a7251f8%2C43799%2C1733777592711:(num 1733777593638) 2024-12-09T20:53:15,294 DEBUG [RS:2;00c23a7251f8:43799 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,294 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:15,294 INFO [RS:2;00c23a7251f8:43799 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T20:53:15,294 INFO [RS:2;00c23a7251f8:43799 {}] hbase.ChoreService(370): Chore service for: regionserver/00c23a7251f8:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-09T20:53:15,294 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T20:53:15,294 INFO [regionserver/00c23a7251f8:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T20:53:15,294 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T20:53:15,294 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T20:53:15,295 INFO [RS:2;00c23a7251f8:43799 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T20:53:15,295 INFO [RS:2;00c23a7251f8:43799 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:43799 2024-12-09T20:53:15,301 INFO [regionserver/00c23a7251f8:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:15,303 INFO [regionserver/00c23a7251f8:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:15,303 INFO [regionserver/00c23a7251f8:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:15,303 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c23a7251f8,43799,1733777592711 2024-12-09T20:53:15,303 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T20:53:15,303 INFO [RS:2;00c23a7251f8:43799 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T20:53:15,305 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/info/6c876b265d404da9ae07d080a4b13cd8 is 153, key is TestHBaseWalOnEC,,1733777594257.10382e18560e2242c4008357cc131575./info:regioninfo/1733777594642/Put/seqid=0 2024-12-09T20:53:15,314 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c23a7251f8,43799,1733777592711] 2024-12-09T20:53:15,314 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741840_1016 (size=6637) 2024-12-09T20:53:15,314 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741840_1016 (size=6637) 2024-12-09T20:53:15,314 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741840_1016 (size=6637) 2024-12-09T20:53:15,315 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/info/6c876b265d404da9ae07d080a4b13cd8 2024-12-09T20:53:15,324 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c23a7251f8,43799,1733777592711 already deleted, retry=false 2024-12-09T20:53:15,324 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c23a7251f8,43799,1733777592711 expired; onlineServers=2 2024-12-09T20:53:15,338 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/ns/eb448d098a88403388dcd2dd370314eb is 43, key is default/ns:d/1733777594123/Put/seqid=0 2024-12-09T20:53:15,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741841_1017 (size=5153) 2024-12-09T20:53:15,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741841_1017 (size=5153) 2024-12-09T20:53:15,345 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741841_1017 (size=5153) 2024-12-09T20:53:15,345 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/ns/eb448d098a88403388dcd2dd370314eb 2024-12-09T20:53:15,367 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/table/916b6260bfc44b129cbca12bca4c8e25 is 52, key is TestHBaseWalOnEC/table:state/1733777594657/Put/seqid=0 2024-12-09T20:53:15,373 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741842_1018 (size=5249) 2024-12-09T20:53:15,374 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741842_1018 (size=5249) 2024-12-09T20:53:15,374 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741842_1018 (size=5249) 2024-12-09T20:53:15,374 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/table/916b6260bfc44b129cbca12bca4c8e25 2024-12-09T20:53:15,382 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/info/6c876b265d404da9ae07d080a4b13cd8 as hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/info/6c876b265d404da9ae07d080a4b13cd8 2024-12-09T20:53:15,390 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/info/6c876b265d404da9ae07d080a4b13cd8, entries=10, sequenceid=11, filesize=6.5 K 2024-12-09T20:53:15,391 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/ns/eb448d098a88403388dcd2dd370314eb as hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/ns/eb448d098a88403388dcd2dd370314eb 2024-12-09T20:53:15,397 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/ns/eb448d098a88403388dcd2dd370314eb, entries=2, sequenceid=11, filesize=5.0 K 2024-12-09T20:53:15,399 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/.tmp/table/916b6260bfc44b129cbca12bca4c8e25 as hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/table/916b6260bfc44b129cbca12bca4c8e25 2024-12-09T20:53:15,405 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/table/916b6260bfc44b129cbca12bca4c8e25, entries=2, sequenceid=11, filesize=5.1 K 2024-12-09T20:53:15,407 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 121ms, sequenceid=11, compaction requested=false 2024-12-09T20:53:15,412 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-09T20:53:15,413 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-09T20:53:15,413 INFO [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-09T20:53:15,413 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733777595286Running coprocessor pre-close hooks at 1733777595286Disabling compacts and flushes for region at 1733777595286Disabling writes for close at 1733777595286Obtaining lock to block concurrent updates at 1733777595286Preparing flush snapshotting stores in 1588230740 at 1733777595286Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733777595287 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733777595287Flushing 1588230740/info: creating writer at 1733777595287Flushing 1588230740/info: appending metadata at 1733777595304 (+17 ms)Flushing 1588230740/info: closing flushed file at 1733777595304Flushing 1588230740/ns: creating writer at 1733777595322 (+18 ms)Flushing 1588230740/ns: appending metadata at 1733777595337 (+15 ms)Flushing 1588230740/ns: closing flushed file at 1733777595337Flushing 1588230740/table: creating writer at 1733777595353 (+16 ms)Flushing 1588230740/table: appending metadata at 1733777595366 (+13 ms)Flushing 1588230740/table: closing flushed file at 1733777595366Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7efd7190: reopening flushed file at 1733777595381 (+15 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4f02be39: reopening flushed file at 1733777595390 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3e04299f: reopening flushed file at 1733777595398 (+8 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 121ms, sequenceid=11, compaction requested=false at 1733777595407 (+9 ms)Writing region close event to WAL at 1733777595408 (+1 ms)Running coprocessor post-close hooks at 1733777595413 (+5 ms)Closed at 1733777595413 2024-12-09T20:53:15,413 DEBUG [RS_CLOSE_META-regionserver/00c23a7251f8:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-09T20:53:15,414 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:15,414 INFO [RS:2;00c23a7251f8:43799 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T20:53:15,414 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43799-0x1000c5a975e0003, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:15,414 INFO [RS:2;00c23a7251f8:43799 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c23a7251f8,43799,1733777592711; zookeeper connection closed. 2024-12-09T20:53:15,414 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@1668905 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@1668905 2024-12-09T20:53:15,486 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(976): stopping server 00c23a7251f8,36855,1733777592682; all regions closed. 2024-12-09T20:53:15,486 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(976): stopping server 00c23a7251f8,36251,1733777592647; all regions closed. 2024-12-09T20:53:15,487 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,487 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741833_1009 (size=1298) 2024-12-09T20:53:15,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741836_1012 (size=2751) 2024-12-09T20:53:15,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741836_1012 (size=2751) 2024-12-09T20:53:15,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741836_1012 (size=2751) 2024-12-09T20:53:15,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741833_1009 (size=1298) 2024-12-09T20:53:15,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741833_1009 (size=1298) 2024-12-09T20:53:15,496 DEBUG [RS:0;00c23a7251f8:36251 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/oldWALs 2024-12-09T20:53:15,496 INFO [RS:0;00c23a7251f8:36251 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 00c23a7251f8%2C36251%2C1733777592647:(num 1733777593638) 2024-12-09T20:53:15,496 DEBUG [RS:0;00c23a7251f8:36251 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,496 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:15,496 DEBUG [RS:1;00c23a7251f8:36855 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/oldWALs 2024-12-09T20:53:15,496 INFO [RS:1;00c23a7251f8:36855 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 00c23a7251f8%2C36855%2C1733777592682.meta:.meta(num 1733777593981) 2024-12-09T20:53:15,496 INFO [RS:0;00c23a7251f8:36251 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T20:53:15,497 INFO [RS:0;00c23a7251f8:36251 {}] hbase.ChoreService(370): Chore service for: regionserver/00c23a7251f8:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-09T20:53:15,497 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T20:53:15,497 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,497 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T20:53:15,497 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T20:53:15,497 INFO [RS:0;00c23a7251f8:36251 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T20:53:15,497 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,497 INFO [RS:0;00c23a7251f8:36251 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36251 2024-12-09T20:53:15,497 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,497 INFO [regionserver/00c23a7251f8:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T20:53:15,497 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,498 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741835_1011 (size=93) 2024-12-09T20:53:15,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741835_1011 (size=93) 2024-12-09T20:53:15,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741835_1011 (size=93) 2024-12-09T20:53:15,501 INFO [regionserver/00c23a7251f8:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-09T20:53:15,501 INFO [regionserver/00c23a7251f8:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-09T20:53:15,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T20:53:15,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c23a7251f8,36251,1733777592647 2024-12-09T20:53:15,503 INFO [RS:0;00c23a7251f8:36251 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T20:53:15,503 DEBUG [RS:1;00c23a7251f8:36855 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/oldWALs 2024-12-09T20:53:15,503 INFO [RS:1;00c23a7251f8:36855 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 00c23a7251f8%2C36855%2C1733777592682:(num 1733777593645) 2024-12-09T20:53:15,503 DEBUG [RS:1;00c23a7251f8:36855 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T20:53:15,503 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T20:53:15,503 INFO [RS:1;00c23a7251f8:36855 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T20:53:15,503 INFO [RS:1;00c23a7251f8:36855 {}] hbase.ChoreService(370): Chore service for: regionserver/00c23a7251f8:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-09T20:53:15,504 INFO [RS:1;00c23a7251f8:36855 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T20:53:15,504 INFO [regionserver/00c23a7251f8:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T20:53:15,504 INFO [RS:1;00c23a7251f8:36855 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36855 2024-12-09T20:53:15,513 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c23a7251f8,36251,1733777592647] 2024-12-09T20:53:15,524 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c23a7251f8,36855,1733777592682 2024-12-09T20:53:15,524 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T20:53:15,524 INFO [RS:1;00c23a7251f8:36855 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T20:53:15,534 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c23a7251f8,36251,1733777592647 already deleted, retry=false 2024-12-09T20:53:15,534 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c23a7251f8,36251,1733777592647 expired; onlineServers=1 2024-12-09T20:53:15,545 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c23a7251f8,36855,1733777592682] 2024-12-09T20:53:15,555 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c23a7251f8,36855,1733777592682 already deleted, retry=false 2024-12-09T20:53:15,555 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c23a7251f8,36855,1733777592682 expired; onlineServers=0 2024-12-09T20:53:15,555 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '00c23a7251f8,38673,1733777592492' ***** 2024-12-09T20:53:15,555 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-09T20:53:15,556 INFO [M:0;00c23a7251f8:38673 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T20:53:15,556 INFO [M:0;00c23a7251f8:38673 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T20:53:15,556 DEBUG [M:0;00c23a7251f8:38673 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-09T20:53:15,556 DEBUG [M:0;00c23a7251f8:38673 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-09T20:53:15,556 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-09T20:53:15,556 DEBUG [master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.large.0-1733777593310 {}] cleaner.HFileCleaner(306): Exit Thread[master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.large.0-1733777593310,5,FailOnTimeoutGroup] 2024-12-09T20:53:15,556 DEBUG [master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.small.0-1733777593310 {}] cleaner.HFileCleaner(306): Exit Thread[master/00c23a7251f8:0:becomeActiveMaster-HFileCleaner.small.0-1733777593310,5,FailOnTimeoutGroup] 2024-12-09T20:53:15,556 INFO [M:0;00c23a7251f8:38673 {}] hbase.ChoreService(370): Chore service for: master/00c23a7251f8:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-09T20:53:15,556 INFO [M:0;00c23a7251f8:38673 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T20:53:15,556 DEBUG [M:0;00c23a7251f8:38673 {}] master.HMaster(1795): Stopping service threads 2024-12-09T20:53:15,556 INFO [M:0;00c23a7251f8:38673 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-09T20:53:15,556 INFO [M:0;00c23a7251f8:38673 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-09T20:53:15,557 INFO [M:0;00c23a7251f8:38673 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-09T20:53:15,557 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-09T20:53:15,566 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-09T20:53:15,566 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T20:53:15,566 DEBUG [M:0;00c23a7251f8:38673 {}] zookeeper.ZKUtil(347): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-09T20:53:15,566 WARN [M:0;00c23a7251f8:38673 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-09T20:53:15,567 INFO [M:0;00c23a7251f8:38673 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/.lastflushedseqids 2024-12-09T20:53:15,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741843_1019 (size=127) 2024-12-09T20:53:15,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741843_1019 (size=127) 2024-12-09T20:53:15,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741843_1019 (size=127) 2024-12-09T20:53:15,576 INFO [M:0;00c23a7251f8:38673 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-09T20:53:15,576 INFO [M:0;00c23a7251f8:38673 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-09T20:53:15,576 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-09T20:53:15,576 INFO [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:15,576 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:15,576 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-09T20:53:15,576 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:15,576 INFO [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.84 KB heapSize=34.13 KB 2024-12-09T20:53:15,592 DEBUG [M:0;00c23a7251f8:38673 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/ea46bafe5fa84b12b30349f0770c70e5 is 82, key is hbase:meta,,1/info:regioninfo/1733777594019/Put/seqid=0 2024-12-09T20:53:15,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741844_1020 (size=5672) 2024-12-09T20:53:15,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741844_1020 (size=5672) 2024-12-09T20:53:15,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741844_1020 (size=5672) 2024-12-09T20:53:15,600 INFO [M:0;00c23a7251f8:38673 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/ea46bafe5fa84b12b30349f0770c70e5 2024-12-09T20:53:15,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:15,614 INFO [RS:0;00c23a7251f8:36251 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T20:53:15,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36251-0x1000c5a975e0001, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:15,614 INFO [RS:0;00c23a7251f8:36251 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c23a7251f8,36251,1733777592647; zookeeper connection closed. 2024-12-09T20:53:15,614 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@6e12f877 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@6e12f877 2024-12-09T20:53:15,628 DEBUG [M:0;00c23a7251f8:38673 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/d9d4a5a83bef4a60858887b80a518031 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733777594664/Put/seqid=0 2024-12-09T20:53:15,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741845_1021 (size=6440) 2024-12-09T20:53:15,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741845_1021 (size=6440) 2024-12-09T20:53:15,636 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741845_1021 (size=6440) 2024-12-09T20:53:15,637 INFO [M:0;00c23a7251f8:38673 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.15 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/d9d4a5a83bef4a60858887b80a518031 2024-12-09T20:53:15,645 INFO [RS:1;00c23a7251f8:36855 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T20:53:15,645 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:15,645 INFO [RS:1;00c23a7251f8:36855 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c23a7251f8,36855,1733777592682; zookeeper connection closed. 2024-12-09T20:53:15,645 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36855-0x1000c5a975e0002, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:15,646 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5dbcfe22 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5dbcfe22 2024-12-09T20:53:15,646 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-09T20:53:15,659 DEBUG [M:0;00c23a7251f8:38673 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/9109a6333e5945e28c24fa0adc03a34f is 69, key is 00c23a7251f8,36251,1733777592647/rs:state/1733777593398/Put/seqid=0 2024-12-09T20:53:15,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741846_1022 (size=5294) 2024-12-09T20:53:15,669 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741846_1022 (size=5294) 2024-12-09T20:53:15,669 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741846_1022 (size=5294) 2024-12-09T20:53:15,669 INFO [M:0;00c23a7251f8:38673 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/9109a6333e5945e28c24fa0adc03a34f 2024-12-09T20:53:15,676 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/ea46bafe5fa84b12b30349f0770c70e5 as hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/ea46bafe5fa84b12b30349f0770c70e5 2024-12-09T20:53:15,685 INFO [M:0;00c23a7251f8:38673 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/ea46bafe5fa84b12b30349f0770c70e5, entries=8, sequenceid=72, filesize=5.5 K 2024-12-09T20:53:15,686 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/d9d4a5a83bef4a60858887b80a518031 as hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/d9d4a5a83bef4a60858887b80a518031 2024-12-09T20:53:15,693 INFO [M:0;00c23a7251f8:38673 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/d9d4a5a83bef4a60858887b80a518031, entries=8, sequenceid=72, filesize=6.3 K 2024-12-09T20:53:15,695 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/9109a6333e5945e28c24fa0adc03a34f as hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/9109a6333e5945e28c24fa0adc03a34f 2024-12-09T20:53:15,702 INFO [M:0;00c23a7251f8:38673 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33191/user/jenkins/test-data/0b30d223-ee52-a55f-2c02-945b20dae7c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/9109a6333e5945e28c24fa0adc03a34f, entries=3, sequenceid=72, filesize=5.2 K 2024-12-09T20:53:15,704 INFO [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 128ms, sequenceid=72, compaction requested=false 2024-12-09T20:53:15,705 INFO [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T20:53:15,705 DEBUG [M:0;00c23a7251f8:38673 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733777595576Disabling compacts and flushes for region at 1733777595576Disabling writes for close at 1733777595576Obtaining lock to block concurrent updates at 1733777595576Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733777595576Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27480, getHeapSize=34880, getOffHeapSize=0, getCellsCount=85 at 1733777595577 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733777595578 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733777595578Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733777595591 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733777595592 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733777595606 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733777595628 (+22 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733777595628Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733777595644 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733777595658 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733777595658Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@30588299: reopening flushed file at 1733777595675 (+17 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@13c13076: reopening flushed file at 1733777595685 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@57dc9d8f: reopening flushed file at 1733777595694 (+9 ms)Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 128ms, sequenceid=72, compaction requested=false at 1733777595704 (+10 ms)Writing region close event to WAL at 1733777595705 (+1 ms)Closed at 1733777595705 2024-12-09T20:53:15,705 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,705 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,706 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,706 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,706 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T20:53:15,708 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35583 is added to blk_1073741830_1006 (size=32683) 2024-12-09T20:53:15,708 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45413 is added to blk_1073741830_1006 (size=32683) 2024-12-09T20:53:15,708 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34569 is added to blk_1073741830_1006 (size=32683) 2024-12-09T20:53:15,709 INFO [M:0;00c23a7251f8:38673 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-09T20:53:15,709 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T20:53:15,709 INFO [M:0;00c23a7251f8:38673 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:38673 2024-12-09T20:53:15,709 INFO [M:0;00c23a7251f8:38673 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T20:53:15,814 INFO [M:0;00c23a7251f8:38673 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T20:53:15,814 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:15,814 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38673-0x1000c5a975e0000, quorum=127.0.0.1:62017, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T20:53:15,817 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@733029a3{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:15,818 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@a87cd5{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T20:53:15,818 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T20:53:15,818 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@60d8940e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T20:53:15,819 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@39179133{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir/,STOPPED} 2024-12-09T20:53:15,820 WARN [BP-216836135-172.17.0.2-1733777590115 heartbeating to localhost/127.0.0.1:33191 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T20:53:15,820 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T20:53:15,820 WARN [BP-216836135-172.17.0.2-1733777590115 heartbeating to localhost/127.0.0.1:33191 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-216836135-172.17.0.2-1733777590115 (Datanode Uuid 6ca5fd4c-c694-47f2-8311-4d5ac60e4c1b) service to localhost/127.0.0.1:33191 2024-12-09T20:53:15,820 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T20:53:15,821 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data5/current/BP-216836135-172.17.0.2-1733777590115 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:15,822 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data6/current/BP-216836135-172.17.0.2-1733777590115 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:15,822 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T20:53:15,825 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4be50faa{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:15,826 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@6625a4f2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T20:53:15,826 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T20:53:15,826 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6f5c60f4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T20:53:15,826 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7febc9c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir/,STOPPED} 2024-12-09T20:53:15,829 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T20:53:15,829 WARN [BP-216836135-172.17.0.2-1733777590115 heartbeating to localhost/127.0.0.1:33191 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T20:53:15,829 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T20:53:15,829 WARN [BP-216836135-172.17.0.2-1733777590115 heartbeating to localhost/127.0.0.1:33191 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-216836135-172.17.0.2-1733777590115 (Datanode Uuid 61260aaa-ea76-4f6e-9008-27f31da5f457) service to localhost/127.0.0.1:33191 2024-12-09T20:53:15,830 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data3/current/BP-216836135-172.17.0.2-1733777590115 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:15,830 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data4/current/BP-216836135-172.17.0.2-1733777590115 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:15,830 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T20:53:15,833 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@765c7210{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T20:53:15,833 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3e498d5c{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T20:53:15,833 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T20:53:15,833 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@46f2e60d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T20:53:15,833 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6decf963{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir/,STOPPED} 2024-12-09T20:53:15,835 WARN [BP-216836135-172.17.0.2-1733777590115 heartbeating to localhost/127.0.0.1:33191 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T20:53:15,835 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T20:53:15,835 WARN [BP-216836135-172.17.0.2-1733777590115 heartbeating to localhost/127.0.0.1:33191 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-216836135-172.17.0.2-1733777590115 (Datanode Uuid c09e6b75-01c4-4ae2-aada-2124c1251ba2) service to localhost/127.0.0.1:33191 2024-12-09T20:53:15,835 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T20:53:15,835 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data1/current/BP-216836135-172.17.0.2-1733777590115 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:15,835 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/cluster_b98c40a9-bb3d-171d-774a-4ab5411eba61/data/data2/current/BP-216836135-172.17.0.2-1733777590115 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T20:53:15,836 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T20:53:15,841 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@563c957f{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-09T20:53:15,842 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@cbd9f23{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T20:53:15,842 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T20:53:15,842 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@59bbe271{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T20:53:15,842 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1aa34083{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/fb7fc66b-53ca-854e-0199-1fbe1357fca4/hadoop.log.dir/,STOPPED} 2024-12-09T20:53:15,850 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-09T20:53:15,874 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-09T20:53:15,881 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=154 (was 91) - Thread LEAK? -, OpenFileDescriptor=518 (was 439) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=208 (was 217), ProcessCount=11 (was 11), AvailableMemoryMB=3696 (was 3891)