2024-12-05 13:43:00,191 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-05 13:43:00,204 main DEBUG Took 0.011395 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-05 13:43:00,205 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-05 13:43:00,205 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-05 13:43:00,206 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-05 13:43:00,208 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,223 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-05 13:43:00,238 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,240 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,241 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,241 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,242 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,242 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,243 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,244 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,245 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,245 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,246 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,246 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,247 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,247 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,248 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,248 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,249 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,249 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,250 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,250 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,250 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,251 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,251 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,252 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 13:43:00,252 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,253 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-05 13:43:00,255 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 13:43:00,256 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-05 13:43:00,257 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-05 13:43:00,258 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-05 13:43:00,260 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-05 13:43:00,260 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-05 13:43:00,269 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-05 13:43:00,272 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-05 13:43:00,274 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-05 13:43:00,274 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-05 13:43:00,275 main DEBUG createAppenders(={Console}) 2024-12-05 13:43:00,275 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-05 13:43:00,276 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-05 13:43:00,276 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-05 13:43:00,277 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-05 13:43:00,277 main DEBUG OutputStream closed 2024-12-05 13:43:00,277 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-05 13:43:00,278 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-05 13:43:00,278 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-05 13:43:00,352 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-05 13:43:00,355 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-05 13:43:00,356 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-05 13:43:00,357 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-05 13:43:00,358 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-05 13:43:00,358 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-05 13:43:00,359 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-05 13:43:00,359 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-05 13:43:00,359 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-05 13:43:00,360 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-05 13:43:00,360 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-05 13:43:00,361 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-05 13:43:00,361 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-05 13:43:00,361 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-05 13:43:00,361 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-05 13:43:00,362 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-05 13:43:00,362 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-05 13:43:00,363 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-05 13:43:00,365 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-05 13:43:00,365 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-05 13:43:00,365 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-05 13:43:00,366 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-05T13:43:00,381 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-05 13:43:00,383 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-05 13:43:00,384 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-05T13:43:00,592 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb 2024-12-05T13:43:00,615 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad, deleteOnExit=true 2024-12-05T13:43:00,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/test.cache.data in system properties and HBase conf 2024-12-05T13:43:00,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.tmp.dir in system properties and HBase conf 2024-12-05T13:43:00,617 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir in system properties and HBase conf 2024-12-05T13:43:00,617 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-05T13:43:00,618 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-05T13:43:00,618 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-05T13:43:00,692 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-05T13:43:00,766 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-05T13:43:00,770 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-05T13:43:00,771 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-05T13:43:00,771 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-05T13:43:00,772 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T13:43:00,772 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-05T13:43:00,772 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-05T13:43:00,773 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T13:43:00,773 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T13:43:00,774 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-05T13:43:00,775 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/nfs.dump.dir in system properties and HBase conf 2024-12-05T13:43:00,775 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/java.io.tmpdir in system properties and HBase conf 2024-12-05T13:43:00,775 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T13:43:00,776 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-05T13:43:00,776 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-05T13:43:01,660 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-05T13:43:01,725 INFO [Time-limited test {}] log.Log(170): Logging initialized @2111ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-05T13:43:01,789 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:01,846 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T13:43:01,863 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T13:43:01,864 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T13:43:01,865 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T13:43:01,877 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:01,880 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@760c69c0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir/,AVAILABLE} 2024-12-05T13:43:01,881 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@ce709a8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T13:43:02,049 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@62d6efd9{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/java.io.tmpdir/jetty-localhost-33821-hadoop-hdfs-3_4_1-tests_jar-_-any-9932769768183853823/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T13:43:02,055 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@353d35a1{HTTP/1.1, (http/1.1)}{localhost:33821} 2024-12-05T13:43:02,056 INFO [Time-limited test {}] server.Server(415): Started @2442ms 2024-12-05T13:43:02,515 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:02,521 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T13:43:02,522 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T13:43:02,523 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T13:43:02,523 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T13:43:02,523 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3a5de9e4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir/,AVAILABLE} 2024-12-05T13:43:02,524 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@69893329{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T13:43:02,617 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1b97a472{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/java.io.tmpdir/jetty-localhost-44319-hadoop-hdfs-3_4_1-tests_jar-_-any-6977115634146631215/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:02,617 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3722a29b{HTTP/1.1, (http/1.1)}{localhost:44319} 2024-12-05T13:43:02,617 INFO [Time-limited test {}] server.Server(415): Started @3004ms 2024-12-05T13:43:02,662 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T13:43:02,755 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:02,761 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T13:43:02,767 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T13:43:02,767 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T13:43:02,767 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T13:43:02,768 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@510fec09{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir/,AVAILABLE} 2024-12-05T13:43:02,769 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@40eb7053{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T13:43:02,874 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@353955e9{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/java.io.tmpdir/jetty-localhost-40293-hadoop-hdfs-3_4_1-tests_jar-_-any-13314006184400682847/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:02,874 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11738cd8{HTTP/1.1, (http/1.1)}{localhost:40293} 2024-12-05T13:43:02,875 INFO [Time-limited test {}] server.Server(415): Started @3261ms 2024-12-05T13:43:02,877 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T13:43:02,909 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:02,915 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T13:43:02,919 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T13:43:02,919 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T13:43:02,919 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T13:43:02,920 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@16cd567f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir/,AVAILABLE} 2024-12-05T13:43:02,921 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5822645a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T13:43:03,014 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3114ae69{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/java.io.tmpdir/jetty-localhost-40721-hadoop-hdfs-3_4_1-tests_jar-_-any-5868599412967650993/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:03,015 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3c70a874{HTTP/1.1, (http/1.1)}{localhost:40721} 2024-12-05T13:43:03,015 INFO [Time-limited test {}] server.Server(415): Started @3401ms 2024-12-05T13:43:03,017 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T13:43:03,674 WARN [Thread-124 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data3/current/BP-75755572-172.17.0.2-1733406181266/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:03,674 WARN [Thread-125 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data4/current/BP-75755572-172.17.0.2-1733406181266/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:03,674 WARN [Thread-122 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data1/current/BP-75755572-172.17.0.2-1733406181266/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:03,674 WARN [Thread-123 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data2/current/BP-75755572-172.17.0.2-1733406181266/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:03,705 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T13:43:03,705 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T13:43:03,744 WARN [Thread-142 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data5/current/BP-75755572-172.17.0.2-1733406181266/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:03,744 WARN [Thread-143 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data6/current/BP-75755572-172.17.0.2-1733406181266/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:03,749 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x8eaa0a5f947759e with lease ID 0xcb6f696d001a5901: Processing first storage report for DS-e3a3dc0c-fe4d-495b-9b7c-add68a5be82b from datanode DatanodeRegistration(127.0.0.1:46105, datanodeUuid=bb905242-3b76-43eb-9622-6d40a1524fde, infoPort=39699, infoSecurePort=0, ipcPort=44259, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266) 2024-12-05T13:43:03,751 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x8eaa0a5f947759e with lease ID 0xcb6f696d001a5901: from storage DS-e3a3dc0c-fe4d-495b-9b7c-add68a5be82b node DatanodeRegistration(127.0.0.1:46105, datanodeUuid=bb905242-3b76-43eb-9622-6d40a1524fde, infoPort=39699, infoSecurePort=0, ipcPort=44259, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2024-12-05T13:43:03,751 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xeb9ce74157f0b80d with lease ID 0xcb6f696d001a5900: Processing first storage report for DS-9f3ba8c7-08c4-4fc4-ab13-d53e34ebb46a from datanode DatanodeRegistration(127.0.0.1:37945, datanodeUuid=3a7446c5-e224-41e9-83cb-7e3e10a1909a, infoPort=36043, infoSecurePort=0, ipcPort=43561, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266) 2024-12-05T13:43:03,752 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xeb9ce74157f0b80d with lease ID 0xcb6f696d001a5900: from storage DS-9f3ba8c7-08c4-4fc4-ab13-d53e34ebb46a node DatanodeRegistration(127.0.0.1:37945, datanodeUuid=3a7446c5-e224-41e9-83cb-7e3e10a1909a, infoPort=36043, infoSecurePort=0, ipcPort=43561, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:03,752 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x8eaa0a5f947759e with lease ID 0xcb6f696d001a5901: Processing first storage report for DS-0e4ade69-2063-4947-b9c1-b96fbcbf988f from datanode DatanodeRegistration(127.0.0.1:46105, datanodeUuid=bb905242-3b76-43eb-9622-6d40a1524fde, infoPort=39699, infoSecurePort=0, ipcPort=44259, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266) 2024-12-05T13:43:03,752 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x8eaa0a5f947759e with lease ID 0xcb6f696d001a5901: from storage DS-0e4ade69-2063-4947-b9c1-b96fbcbf988f node DatanodeRegistration(127.0.0.1:46105, datanodeUuid=bb905242-3b76-43eb-9622-6d40a1524fde, infoPort=39699, infoSecurePort=0, ipcPort=44259, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:03,752 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xeb9ce74157f0b80d with lease ID 0xcb6f696d001a5900: Processing first storage report for DS-6c4d3023-7e0a-41de-935d-06961185c889 from datanode DatanodeRegistration(127.0.0.1:37945, datanodeUuid=3a7446c5-e224-41e9-83cb-7e3e10a1909a, infoPort=36043, infoSecurePort=0, ipcPort=43561, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266) 2024-12-05T13:43:03,753 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xeb9ce74157f0b80d with lease ID 0xcb6f696d001a5900: from storage DS-6c4d3023-7e0a-41de-935d-06961185c889 node DatanodeRegistration(127.0.0.1:37945, datanodeUuid=3a7446c5-e224-41e9-83cb-7e3e10a1909a, infoPort=36043, infoSecurePort=0, ipcPort=43561, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:03,771 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T13:43:03,776 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd18a45140befdbf with lease ID 0xcb6f696d001a5902: Processing first storage report for DS-879f2c03-80b6-4234-87ad-a05ca676da10 from datanode DatanodeRegistration(127.0.0.1:43151, datanodeUuid=1b53e8ef-7425-4750-a0a5-25afca90c6e6, infoPort=34001, infoSecurePort=0, ipcPort=45027, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266) 2024-12-05T13:43:03,777 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd18a45140befdbf with lease ID 0xcb6f696d001a5902: from storage DS-879f2c03-80b6-4234-87ad-a05ca676da10 node DatanodeRegistration(127.0.0.1:43151, datanodeUuid=1b53e8ef-7425-4750-a0a5-25afca90c6e6, infoPort=34001, infoSecurePort=0, ipcPort=45027, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-05T13:43:03,777 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd18a45140befdbf with lease ID 0xcb6f696d001a5902: Processing first storage report for DS-f335f29b-c2fd-4a9e-8b6e-bf1be9f7a096 from datanode DatanodeRegistration(127.0.0.1:43151, datanodeUuid=1b53e8ef-7425-4750-a0a5-25afca90c6e6, infoPort=34001, infoSecurePort=0, ipcPort=45027, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266) 2024-12-05T13:43:03,777 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd18a45140befdbf with lease ID 0xcb6f696d001a5902: from storage DS-f335f29b-c2fd-4a9e-8b6e-bf1be9f7a096 node DatanodeRegistration(127.0.0.1:43151, datanodeUuid=1b53e8ef-7425-4750-a0a5-25afca90c6e6, infoPort=34001, infoSecurePort=0, ipcPort=45027, storageInfo=lv=-57;cid=testClusterID;nsid=1677469881;c=1733406181266), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:03,805 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb 2024-12-05T13:43:03,871 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-05T13:43:03,918 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=162, OpenFileDescriptor=391, MaxFileDescriptor=1048576, SystemLoadAverage=263, ProcessCount=11, AvailableMemoryMB=8422 2024-12-05T13:43:03,920 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-05T13:43:03,927 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-05T13:43:03,998 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/zookeeper_0, clientPort=59506, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-05T13:43:04,008 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=59506 2024-12-05T13:43:04,020 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:04,023 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:04,126 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:04,127 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:04,168 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:59268 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:37945:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59268 dst: /127.0.0.1:37945 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:04,189 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-05T13:43:04,588 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:04,601 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130 with version=8 2024-12-05T13:43:04,602 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/hbase-staging 2024-12-05T13:43:04,683 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-05T13:43:04,900 INFO [Time-limited test {}] client.ConnectionUtils(128): master/da6aa2204f50:0 server-side Connection retries=45 2024-12-05T13:43:04,908 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:04,909 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:04,913 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T13:43:04,913 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:04,913 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T13:43:05,033 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-05T13:43:05,088 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-05T13:43:05,096 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-05T13:43:05,099 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T13:43:05,120 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 64651 (auto-detected) 2024-12-05T13:43:05,121 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-05T13:43:05,136 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36439 2024-12-05T13:43:05,154 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:36439 connecting to ZooKeeper ensemble=127.0.0.1:59506 2024-12-05T13:43:05,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:364390x0, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T13:43:05,235 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:36439-0x101a704f4a10000 connected 2024-12-05T13:43:05,317 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,322 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,331 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:05,335 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130, hbase.cluster.distributed=false 2024-12-05T13:43:05,357 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T13:43:05,361 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36439 2024-12-05T13:43:05,362 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36439 2024-12-05T13:43:05,362 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36439 2024-12-05T13:43:05,362 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36439 2024-12-05T13:43:05,363 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36439 2024-12-05T13:43:05,457 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/da6aa2204f50:0 server-side Connection retries=45 2024-12-05T13:43:05,458 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,459 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,459 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T13:43:05,459 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,459 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T13:43:05,461 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T13:43:05,464 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T13:43:05,464 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:45437 2024-12-05T13:43:05,466 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:45437 connecting to ZooKeeper ensemble=127.0.0.1:59506 2024-12-05T13:43:05,467 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,471 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,487 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:454370x0, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T13:43:05,488 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:454370x0, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:05,488 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45437-0x101a704f4a10001 connected 2024-12-05T13:43:05,492 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T13:43:05,498 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T13:43:05,501 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T13:43:05,505 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T13:43:05,506 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45437 2024-12-05T13:43:05,507 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45437 2024-12-05T13:43:05,507 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45437 2024-12-05T13:43:05,508 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45437 2024-12-05T13:43:05,508 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45437 2024-12-05T13:43:05,522 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/da6aa2204f50:0 server-side Connection retries=45 2024-12-05T13:43:05,523 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,523 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,523 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T13:43:05,523 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,523 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T13:43:05,524 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T13:43:05,524 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T13:43:05,525 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39027 2024-12-05T13:43:05,526 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39027 connecting to ZooKeeper ensemble=127.0.0.1:59506 2024-12-05T13:43:05,527 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,529 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:390270x0, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T13:43:05,540 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:390270x0, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:05,540 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39027-0x101a704f4a10002 connected 2024-12-05T13:43:05,540 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T13:43:05,541 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T13:43:05,542 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T13:43:05,544 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T13:43:05,545 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39027 2024-12-05T13:43:05,546 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39027 2024-12-05T13:43:05,546 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39027 2024-12-05T13:43:05,549 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39027 2024-12-05T13:43:05,549 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39027 2024-12-05T13:43:05,565 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/da6aa2204f50:0 server-side Connection retries=45 2024-12-05T13:43:05,565 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,565 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,566 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T13:43:05,566 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:05,566 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T13:43:05,566 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T13:43:05,566 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T13:43:05,567 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:34025 2024-12-05T13:43:05,569 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:34025 connecting to ZooKeeper ensemble=127.0.0.1:59506 2024-12-05T13:43:05,570 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,572 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,580 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:340250x0, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T13:43:05,581 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:340250x0, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:05,581 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:34025-0x101a704f4a10003 connected 2024-12-05T13:43:05,582 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T13:43:05,583 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T13:43:05,584 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T13:43:05,585 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T13:43:05,586 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34025 2024-12-05T13:43:05,586 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34025 2024-12-05T13:43:05,587 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34025 2024-12-05T13:43:05,587 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34025 2024-12-05T13:43:05,588 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34025 2024-12-05T13:43:05,604 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;da6aa2204f50:36439 2024-12-05T13:43:05,605 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/da6aa2204f50,36439,1733406184749 2024-12-05T13:43:05,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:05,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:05,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:05,614 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:05,617 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/da6aa2204f50,36439,1733406184749 2024-12-05T13:43:05,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T13:43:05,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T13:43:05,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:05,647 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T13:43:05,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:05,647 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:05,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:05,649 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T13:43:05,650 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/da6aa2204f50,36439,1733406184749 from backup master directory 2024-12-05T13:43:05,662 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:05,662 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/da6aa2204f50,36439,1733406184749 2024-12-05T13:43:05,662 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:05,662 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:05,662 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:05,663 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T13:43:05,663 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=da6aa2204f50,36439,1733406184749 2024-12-05T13:43:05,665 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-05T13:43:05,666 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-05T13:43:05,724 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/hbase.id] with ID: d691aebf-c218-40fc-8a5f-118a6447e57f 2024-12-05T13:43:05,724 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/.tmp/hbase.id 2024-12-05T13:43:05,731 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:05,731 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:05,734 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:51752 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:46105:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51752 dst: /127.0.0.1:46105 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:05,740 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-05T13:43:05,741 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:05,741 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/.tmp/hbase.id]:[hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/hbase.id] 2024-12-05T13:43:05,786 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:05,790 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-05T13:43:05,806 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 14ms. 2024-12-05T13:43:05,814 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:05,814 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:05,814 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:05,814 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:05,825 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:05,826 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:05,829 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:49362 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:43151:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49362 dst: /127.0.0.1:43151 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:05,834 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-05T13:43:05,835 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:05,850 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T13:43:05,852 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-05T13:43:05,858 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-05T13:43:05,882 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:05,883 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:05,886 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:51778 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:46105:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51778 dst: /127.0.0.1:46105 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:05,891 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-05T13:43:05,892 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:05,909 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store 2024-12-05T13:43:05,927 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:05,927 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:05,930 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:51792 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:46105:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51792 dst: /127.0.0.1:46105 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:05,935 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-05T13:43:05,936 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:05,939 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-05T13:43:05,942 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:05,943 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T13:43:05,943 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:05,943 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:05,945 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T13:43:05,945 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:05,945 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:05,946 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733406185943Disabling compacts and flushes for region at 1733406185943Disabling writes for close at 1733406185945 (+2 ms)Writing region close event to WAL at 1733406185945Closed at 1733406185945 2024-12-05T13:43:05,948 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/.initializing 2024-12-05T13:43:05,948 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/WALs/da6aa2204f50,36439,1733406184749 2024-12-05T13:43:05,956 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-05T13:43:05,970 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C36439%2C1733406184749, suffix=, logDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/WALs/da6aa2204f50,36439,1733406184749, archiveDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/oldWALs, maxLogs=10 2024-12-05T13:43:05,995 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/WALs/da6aa2204f50,36439,1733406184749/da6aa2204f50%2C36439%2C1733406184749.1733406185974, exclude list is [], retry=0 2024-12-05T13:43:06,015 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:06,016 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46105,DS-e3a3dc0c-fe4d-495b-9b7c-add68a5be82b,DISK] 2024-12-05T13:43:06,016 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37945,DS-9f3ba8c7-08c4-4fc4-ab13-d53e34ebb46a,DISK] 2024-12-05T13:43:06,016 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43151,DS-879f2c03-80b6-4234-87ad-a05ca676da10,DISK] 2024-12-05T13:43:06,019 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-05T13:43:06,057 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/WALs/da6aa2204f50,36439,1733406184749/da6aa2204f50%2C36439%2C1733406184749.1733406185974 2024-12-05T13:43:06,058 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36043:36043),(127.0.0.1/127.0.0.1:34001:34001),(127.0.0.1/127.0.0.1:39699:39699)] 2024-12-05T13:43:06,059 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-05T13:43:06,059 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:06,061 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,062 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,094 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,115 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-05T13:43:06,118 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,120 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:06,120 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,124 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-05T13:43:06,124 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,125 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T13:43:06,126 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,128 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-05T13:43:06,129 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,130 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T13:43:06,130 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,133 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-05T13:43:06,133 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,134 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T13:43:06,134 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,138 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,139 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,144 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,144 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,148 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T13:43:06,151 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:06,157 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T13:43:06,158 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=75261137, jitterRate=0.12147833406925201}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T13:43:06,163 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733406186073Initializing all the Stores at 1733406186075 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406186075Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406186077 (+2 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406186077Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406186077Cleaning up temporary data from old regions at 1733406186144 (+67 ms)Region opened successfully at 1733406186163 (+19 ms) 2024-12-05T13:43:06,164 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-05T13:43:06,194 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7ad653c6, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=da6aa2204f50/172.17.0.2:0 2024-12-05T13:43:06,223 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-05T13:43:06,232 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-05T13:43:06,232 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-05T13:43:06,235 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-05T13:43:06,236 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-05T13:43:06,240 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 4 msec 2024-12-05T13:43:06,240 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-05T13:43:06,263 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-05T13:43:06,271 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-05T13:43:06,305 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-05T13:43:06,309 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-05T13:43:06,312 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-05T13:43:06,322 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-05T13:43:06,325 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-05T13:43:06,329 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-05T13:43:06,338 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-05T13:43:06,340 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-05T13:43:06,347 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-05T13:43:06,366 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-05T13:43:06,372 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-05T13:43:06,380 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:06,380 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:06,380 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:06,380 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:06,380 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,380 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,381 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,381 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,384 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=da6aa2204f50,36439,1733406184749, sessionid=0x101a704f4a10000, setting cluster-up flag (Was=false) 2024-12-05T13:43:06,414 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,414 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,414 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,414 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,439 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-05T13:43:06,443 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=da6aa2204f50,36439,1733406184749 2024-12-05T13:43:06,464 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,464 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,464 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,464 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:06,489 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-05T13:43:06,493 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=da6aa2204f50,36439,1733406184749 2024-12-05T13:43:06,504 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-05T13:43:06,566 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-05T13:43:06,574 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-05T13:43:06,580 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-05T13:43:06,585 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: da6aa2204f50,36439,1733406184749 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-05T13:43:06,591 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(746): ClusterId : d691aebf-c218-40fc-8a5f-118a6447e57f 2024-12-05T13:43:06,591 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(746): ClusterId : d691aebf-c218-40fc-8a5f-118a6447e57f 2024-12-05T13:43:06,591 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(746): ClusterId : d691aebf-c218-40fc-8a5f-118a6447e57f 2024-12-05T13:43:06,592 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/da6aa2204f50:0, corePoolSize=5, maxPoolSize=5 2024-12-05T13:43:06,592 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/da6aa2204f50:0, corePoolSize=5, maxPoolSize=5 2024-12-05T13:43:06,592 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/da6aa2204f50:0, corePoolSize=5, maxPoolSize=5 2024-12-05T13:43:06,592 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/da6aa2204f50:0, corePoolSize=5, maxPoolSize=5 2024-12-05T13:43:06,592 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/da6aa2204f50:0, corePoolSize=10, maxPoolSize=10 2024-12-05T13:43:06,593 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,593 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/da6aa2204f50:0, corePoolSize=2, maxPoolSize=2 2024-12-05T13:43:06,593 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,594 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733406216594 2024-12-05T13:43:06,594 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T13:43:06,594 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T13:43:06,594 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T13:43:06,595 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-05T13:43:06,596 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-05T13:43:06,598 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-05T13:43:06,599 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-05T13:43:06,600 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-05T13:43:06,600 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-05T13:43:06,601 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-05T13:43:06,601 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-05T13:43:06,602 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,605 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,605 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-05T13:43:06,605 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T13:43:06,607 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-05T13:43:06,607 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-05T13:43:06,608 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T13:43:06,608 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T13:43:06,608 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T13:43:06,608 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T13:43:06,609 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T13:43:06,609 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T13:43:06,610 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-05T13:43:06,610 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-05T13:43:06,614 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.large.0-1733406186611,5,FailOnTimeoutGroup] 2024-12-05T13:43:06,614 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.small.0-1733406186614,5,FailOnTimeoutGroup] 2024-12-05T13:43:06,614 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,615 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-05T13:43:06,616 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:06,616 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:06,616 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,617 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,619 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:59304 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:37945:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59304 dst: /127.0.0.1:37945 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:06,623 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T13:43:06,623 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T13:43:06,623 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T13:43:06,624 DEBUG [RS:2;da6aa2204f50:34025 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5d65a0ad, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=da6aa2204f50/172.17.0.2:0 2024-12-05T13:43:06,624 DEBUG [RS:1;da6aa2204f50:39027 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@393ee17, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=da6aa2204f50/172.17.0.2:0 2024-12-05T13:43:06,624 DEBUG [RS:0;da6aa2204f50:45437 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@414cc1e7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=da6aa2204f50/172.17.0.2:0 2024-12-05T13:43:06,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-05T13:43:06,640 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:06,641 DEBUG [RS:1;da6aa2204f50:39027 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;da6aa2204f50:39027 2024-12-05T13:43:06,642 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-05T13:43:06,642 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130 2024-12-05T13:43:06,643 DEBUG [RS:0;da6aa2204f50:45437 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;da6aa2204f50:45437 2024-12-05T13:43:06,644 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;da6aa2204f50:34025 2024-12-05T13:43:06,645 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-05T13:43:06,645 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-05T13:43:06,645 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-05T13:43:06,645 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-05T13:43:06,645 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-05T13:43:06,645 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-05T13:43:06,645 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-05T13:43:06,645 DEBUG [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-05T13:43:06,645 DEBUG [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-05T13:43:06,648 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(2659): reportForDuty to master=da6aa2204f50,36439,1733406184749 with port=45437, startcode=1733406185425 2024-12-05T13:43:06,648 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(2659): reportForDuty to master=da6aa2204f50,36439,1733406184749 with port=34025, startcode=1733406185564 2024-12-05T13:43:06,648 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(2659): reportForDuty to master=da6aa2204f50,36439,1733406184749 with port=39027, startcode=1733406185522 2024-12-05T13:43:06,657 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:06,658 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:06,660 DEBUG [RS:0;da6aa2204f50:45437 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T13:43:06,660 DEBUG [RS:2;da6aa2204f50:34025 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T13:43:06,660 DEBUG [RS:1;da6aa2204f50:39027 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T13:43:06,663 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:59326 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:37945:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59326 dst: /127.0.0.1:37945 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:06,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-05T13:43:06,671 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:06,673 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:06,678 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T13:43:06,681 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T13:43:06,681 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,682 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:06,683 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-05T13:43:06,688 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-05T13:43:06,689 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,690 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:06,690 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T13:43:06,694 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T13:43:06,694 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,695 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:06,695 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T13:43:06,698 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T13:43:06,698 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:06,698 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54435, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T13:43:06,698 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41833, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T13:43:06,698 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49041, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T13:43:06,699 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:06,700 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-05T13:43:06,701 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740 2024-12-05T13:43:06,702 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740 2024-12-05T13:43:06,705 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36439 {}] master.ServerManager(363): Checking decommissioned status of RegionServer da6aa2204f50,34025,1733406185564 2024-12-05T13:43:06,706 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-05T13:43:06,706 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-05T13:43:06,707 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T13:43:06,708 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36439 {}] master.ServerManager(517): Registering regionserver=da6aa2204f50,34025,1733406185564 2024-12-05T13:43:06,711 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-05T13:43:06,719 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36439 {}] master.ServerManager(363): Checking decommissioned status of RegionServer da6aa2204f50,39027,1733406185522 2024-12-05T13:43:06,719 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36439 {}] master.ServerManager(517): Registering regionserver=da6aa2204f50,39027,1733406185522 2024-12-05T13:43:06,723 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130 2024-12-05T13:43:06,723 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:39225 2024-12-05T13:43:06,724 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-05T13:43:06,724 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T13:43:06,725 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36439 {}] master.ServerManager(363): Checking decommissioned status of RegionServer da6aa2204f50,45437,1733406185425 2024-12-05T13:43:06,726 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36439 {}] master.ServerManager(517): Registering regionserver=da6aa2204f50,45437,1733406185425 2024-12-05T13:43:06,726 DEBUG [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130 2024-12-05T13:43:06,726 DEBUG [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:39225 2024-12-05T13:43:06,726 DEBUG [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-05T13:43:06,729 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65061237, jitterRate=-0.030512019991874695}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T13:43:06,731 DEBUG [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130 2024-12-05T13:43:06,731 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733406186673Initializing all the Stores at 1733406186676 (+3 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406186676Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406186677 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406186678 (+1 ms)Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406186678Cleaning up temporary data from old regions at 1733406186706 (+28 ms)Region opened successfully at 1733406186731 (+25 ms) 2024-12-05T13:43:06,731 DEBUG [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:39225 2024-12-05T13:43:06,731 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-05T13:43:06,731 DEBUG [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-05T13:43:06,731 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-05T13:43:06,731 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-05T13:43:06,731 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T13:43:06,731 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T13:43:06,733 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-05T13:43:06,733 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733406186731Disabling compacts and flushes for region at 1733406186731Disabling writes for close at 1733406186731Writing region close event to WAL at 1733406186732 (+1 ms)Closed at 1733406186733 (+1 ms) 2024-12-05T13:43:06,736 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-05T13:43:06,736 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-05T13:43:06,737 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T13:43:06,743 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-05T13:43:06,751 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T13:43:06,753 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-05T13:43:06,754 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-05T13:43:06,755 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-05T13:43:06,765 DEBUG [RS:1;da6aa2204f50:39027 {}] zookeeper.ZKUtil(111): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/da6aa2204f50,39027,1733406185522 2024-12-05T13:43:06,766 WARN [RS:1;da6aa2204f50:39027 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T13:43:06,766 INFO [RS:1;da6aa2204f50:39027 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-05T13:43:06,766 DEBUG [RS:2;da6aa2204f50:34025 {}] zookeeper.ZKUtil(111): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/da6aa2204f50,34025,1733406185564 2024-12-05T13:43:06,766 DEBUG [RS:0;da6aa2204f50:45437 {}] zookeeper.ZKUtil(111): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/da6aa2204f50,45437,1733406185425 2024-12-05T13:43:06,766 DEBUG [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,39027,1733406185522 2024-12-05T13:43:06,766 WARN [RS:2;da6aa2204f50:34025 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T13:43:06,766 WARN [RS:0;da6aa2204f50:45437 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T13:43:06,766 INFO [RS:2;da6aa2204f50:34025 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-05T13:43:06,766 INFO [RS:0;da6aa2204f50:45437 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-05T13:43:06,766 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,34025,1733406185564 2024-12-05T13:43:06,766 DEBUG [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,45437,1733406185425 2024-12-05T13:43:06,768 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [da6aa2204f50,45437,1733406185425] 2024-12-05T13:43:06,768 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [da6aa2204f50,39027,1733406185522] 2024-12-05T13:43:06,768 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [da6aa2204f50,34025,1733406185564] 2024-12-05T13:43:06,788 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T13:43:06,788 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T13:43:06,788 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T13:43:06,803 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T13:43:06,803 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T13:43:06,803 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T13:43:06,808 INFO [RS:1;da6aa2204f50:39027 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T13:43:06,808 INFO [RS:2;da6aa2204f50:34025 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T13:43:06,808 INFO [RS:0;da6aa2204f50:45437 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T13:43:06,808 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,808 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,808 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,809 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-05T13:43:06,809 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-05T13:43:06,809 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-05T13:43:06,815 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-05T13:43:06,815 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-05T13:43:06,815 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-05T13:43:06,816 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,816 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,816 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,816 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,816 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/da6aa2204f50:0, corePoolSize=2, maxPoolSize=2 2024-12-05T13:43:06,817 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/da6aa2204f50:0, corePoolSize=2, maxPoolSize=2 2024-12-05T13:43:06,817 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/da6aa2204f50:0, corePoolSize=2, maxPoolSize=2 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:06,817 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:0;da6aa2204f50:45437 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:06,817 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,817 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,818 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,818 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,818 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,818 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:06,818 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:06,818 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:06,818 DEBUG [RS:1;da6aa2204f50:39027 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:06,818 DEBUG [RS:2;da6aa2204f50:34025 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:06,818 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,819 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,819 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,819 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,819 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,819 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,45437,1733406185425-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T13:43:06,820 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,820 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,820 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,820 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,820 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,820 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,34025,1733406185564-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T13:43:06,821 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,821 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,821 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,821 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,821 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,821 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,39027,1733406185522-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T13:43:06,837 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T13:43:06,838 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,34025,1733406185564-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,839 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,839 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.Replication(171): da6aa2204f50,34025,1733406185564 started 2024-12-05T13:43:06,840 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T13:43:06,840 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,45437,1733406185425-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,840 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,840 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.Replication(171): da6aa2204f50,45437,1733406185425 started 2024-12-05T13:43:06,842 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T13:43:06,842 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,39027,1733406185522-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,842 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,843 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.Replication(171): da6aa2204f50,39027,1733406185522 started 2024-12-05T13:43:06,857 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,858 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1482): Serving as da6aa2204f50,34025,1733406185564, RpcServer on da6aa2204f50/172.17.0.2:34025, sessionid=0x101a704f4a10003 2024-12-05T13:43:06,859 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T13:43:06,859 DEBUG [RS:2;da6aa2204f50:34025 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager da6aa2204f50,34025,1733406185564 2024-12-05T13:43:06,859 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,34025,1733406185564' 2024-12-05T13:43:06,859 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T13:43:06,860 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T13:43:06,861 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,861 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(1482): Serving as da6aa2204f50,45437,1733406185425, RpcServer on da6aa2204f50/172.17.0.2:45437, sessionid=0x101a704f4a10001 2024-12-05T13:43:06,861 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T13:43:06,861 DEBUG [RS:0;da6aa2204f50:45437 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager da6aa2204f50,45437,1733406185425 2024-12-05T13:43:06,861 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,45437,1733406185425' 2024-12-05T13:43:06,861 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T13:43:06,861 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T13:43:06,861 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T13:43:06,861 DEBUG [RS:2;da6aa2204f50:34025 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager da6aa2204f50,34025,1733406185564 2024-12-05T13:43:06,861 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,34025,1733406185564' 2024-12-05T13:43:06,862 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T13:43:06,862 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T13:43:06,862 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T13:43:06,862 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T13:43:06,862 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T13:43:06,863 DEBUG [RS:0;da6aa2204f50:45437 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager da6aa2204f50,45437,1733406185425 2024-12-05T13:43:06,863 DEBUG [RS:2;da6aa2204f50:34025 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T13:43:06,863 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,45437,1733406185425' 2024-12-05T13:43:06,863 INFO [RS:2;da6aa2204f50:34025 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T13:43:06,863 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T13:43:06,863 INFO [RS:2;da6aa2204f50:34025 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T13:43:06,863 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T13:43:06,863 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:06,864 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1482): Serving as da6aa2204f50,39027,1733406185522, RpcServer on da6aa2204f50/172.17.0.2:39027, sessionid=0x101a704f4a10002 2024-12-05T13:43:06,864 DEBUG [RS:0;da6aa2204f50:45437 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T13:43:06,864 INFO [RS:0;da6aa2204f50:45437 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T13:43:06,864 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T13:43:06,864 INFO [RS:0;da6aa2204f50:45437 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T13:43:06,864 DEBUG [RS:1;da6aa2204f50:39027 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager da6aa2204f50,39027,1733406185522 2024-12-05T13:43:06,864 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,39027,1733406185522' 2024-12-05T13:43:06,864 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T13:43:06,864 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T13:43:06,865 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T13:43:06,865 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T13:43:06,865 DEBUG [RS:1;da6aa2204f50:39027 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager da6aa2204f50,39027,1733406185522 2024-12-05T13:43:06,865 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,39027,1733406185522' 2024-12-05T13:43:06,865 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T13:43:06,866 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T13:43:06,866 DEBUG [RS:1;da6aa2204f50:39027 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T13:43:06,866 INFO [RS:1;da6aa2204f50:39027 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T13:43:06,866 INFO [RS:1;da6aa2204f50:39027 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T13:43:06,906 WARN [da6aa2204f50:36439 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-05T13:43:06,968 INFO [RS:2;da6aa2204f50:34025 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-05T13:43:06,969 INFO [RS:1;da6aa2204f50:39027 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-05T13:43:06,969 INFO [RS:0;da6aa2204f50:45437 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-05T13:43:06,972 INFO [RS:0;da6aa2204f50:45437 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C45437%2C1733406185425, suffix=, logDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,45437,1733406185425, archiveDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/oldWALs, maxLogs=32 2024-12-05T13:43:06,972 INFO [RS:2;da6aa2204f50:34025 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C34025%2C1733406185564, suffix=, logDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,34025,1733406185564, archiveDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/oldWALs, maxLogs=32 2024-12-05T13:43:06,972 INFO [RS:1;da6aa2204f50:39027 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C39027%2C1733406185522, suffix=, logDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,39027,1733406185522, archiveDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/oldWALs, maxLogs=32 2024-12-05T13:43:06,989 DEBUG [RS:0;da6aa2204f50:45437 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,45437,1733406185425/da6aa2204f50%2C45437%2C1733406185425.1733406186977, exclude list is [], retry=0 2024-12-05T13:43:06,989 DEBUG [RS:2;da6aa2204f50:34025 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,34025,1733406185564/da6aa2204f50%2C34025%2C1733406185564.1733406186977, exclude list is [], retry=0 2024-12-05T13:43:06,989 DEBUG [RS:1;da6aa2204f50:39027 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,39027,1733406185522/da6aa2204f50%2C39027%2C1733406185522.1733406186977, exclude list is [], retry=0 2024-12-05T13:43:06,994 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43151,DS-879f2c03-80b6-4234-87ad-a05ca676da10,DISK] 2024-12-05T13:43:06,994 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46105,DS-e3a3dc0c-fe4d-495b-9b7c-add68a5be82b,DISK] 2024-12-05T13:43:06,994 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37945,DS-9f3ba8c7-08c4-4fc4-ab13-d53e34ebb46a,DISK] 2024-12-05T13:43:06,995 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46105,DS-e3a3dc0c-fe4d-495b-9b7c-add68a5be82b,DISK] 2024-12-05T13:43:06,995 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43151,DS-879f2c03-80b6-4234-87ad-a05ca676da10,DISK] 2024-12-05T13:43:06,996 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46105,DS-e3a3dc0c-fe4d-495b-9b7c-add68a5be82b,DISK] 2024-12-05T13:43:06,996 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37945,DS-9f3ba8c7-08c4-4fc4-ab13-d53e34ebb46a,DISK] 2024-12-05T13:43:07,027 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43151,DS-879f2c03-80b6-4234-87ad-a05ca676da10,DISK] 2024-12-05T13:43:07,027 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37945,DS-9f3ba8c7-08c4-4fc4-ab13-d53e34ebb46a,DISK] 2024-12-05T13:43:07,033 INFO [RS:0;da6aa2204f50:45437 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,45437,1733406185425/da6aa2204f50%2C45437%2C1733406185425.1733406186977 2024-12-05T13:43:07,034 DEBUG [RS:0;da6aa2204f50:45437 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34001:34001),(127.0.0.1/127.0.0.1:39699:39699),(127.0.0.1/127.0.0.1:36043:36043)] 2024-12-05T13:43:07,034 INFO [RS:2;da6aa2204f50:34025 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,34025,1733406185564/da6aa2204f50%2C34025%2C1733406185564.1733406186977 2024-12-05T13:43:07,035 DEBUG [RS:2;da6aa2204f50:34025 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36043:36043),(127.0.0.1/127.0.0.1:39699:39699),(127.0.0.1/127.0.0.1:34001:34001)] 2024-12-05T13:43:07,035 INFO [RS:1;da6aa2204f50:39027 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,39027,1733406185522/da6aa2204f50%2C39027%2C1733406185522.1733406186977 2024-12-05T13:43:07,035 DEBUG [RS:1;da6aa2204f50:39027 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:39699:39699),(127.0.0.1/127.0.0.1:34001:34001),(127.0.0.1/127.0.0.1:36043:36043)] 2024-12-05T13:43:07,159 DEBUG [da6aa2204f50:36439 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-05T13:43:07,168 DEBUG [da6aa2204f50:36439 {}] balancer.BalancerClusterState(204): Hosts are {da6aa2204f50=0} racks are {/default-rack=0} 2024-12-05T13:43:07,174 DEBUG [da6aa2204f50:36439 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-05T13:43:07,174 DEBUG [da6aa2204f50:36439 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-05T13:43:07,175 DEBUG [da6aa2204f50:36439 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-05T13:43:07,175 DEBUG [da6aa2204f50:36439 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-05T13:43:07,175 DEBUG [da6aa2204f50:36439 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-05T13:43:07,175 DEBUG [da6aa2204f50:36439 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-05T13:43:07,175 INFO [da6aa2204f50:36439 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-05T13:43:07,175 INFO [da6aa2204f50:36439 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-05T13:43:07,175 INFO [da6aa2204f50:36439 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-05T13:43:07,175 DEBUG [da6aa2204f50:36439 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-05T13:43:07,181 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=da6aa2204f50,34025,1733406185564 2024-12-05T13:43:07,187 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as da6aa2204f50,34025,1733406185564, state=OPENING 2024-12-05T13:43:07,230 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-05T13:43:07,239 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:07,239 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:07,239 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:07,239 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:07,239 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:07,239 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:07,239 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:07,240 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:07,241 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T13:43:07,243 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=da6aa2204f50,34025,1733406185564}] 2024-12-05T13:43:07,420 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T13:43:07,422 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:32919, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T13:43:07,436 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-05T13:43:07,437 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-05T13:43:07,437 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-05T13:43:07,440 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C34025%2C1733406185564.meta, suffix=.meta, logDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,34025,1733406185564, archiveDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/oldWALs, maxLogs=32 2024-12-05T13:43:07,453 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,34025,1733406185564/da6aa2204f50%2C34025%2C1733406185564.meta.1733406187442.meta, exclude list is [], retry=0 2024-12-05T13:43:07,458 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43151,DS-879f2c03-80b6-4234-87ad-a05ca676da10,DISK] 2024-12-05T13:43:07,458 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46105,DS-e3a3dc0c-fe4d-495b-9b7c-add68a5be82b,DISK] 2024-12-05T13:43:07,458 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37945,DS-9f3ba8c7-08c4-4fc4-ab13-d53e34ebb46a,DISK] 2024-12-05T13:43:07,461 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,34025,1733406185564/da6aa2204f50%2C34025%2C1733406185564.meta.1733406187442.meta 2024-12-05T13:43:07,461 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34001:34001),(127.0.0.1/127.0.0.1:39699:39699),(127.0.0.1/127.0.0.1:36043:36043)] 2024-12-05T13:43:07,461 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-05T13:43:07,463 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-05T13:43:07,465 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-05T13:43:07,470 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-05T13:43:07,473 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-05T13:43:07,474 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:07,474 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-05T13:43:07,474 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-05T13:43:07,477 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T13:43:07,479 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T13:43:07,479 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:07,480 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:07,480 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-05T13:43:07,482 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-05T13:43:07,482 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:07,483 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:07,483 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T13:43:07,485 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T13:43:07,485 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:07,486 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:07,486 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T13:43:07,487 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T13:43:07,487 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:07,488 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:07,488 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-05T13:43:07,490 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740 2024-12-05T13:43:07,493 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740 2024-12-05T13:43:07,495 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-05T13:43:07,495 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-05T13:43:07,496 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T13:43:07,499 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-05T13:43:07,500 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61708030, jitterRate=-0.08047869801521301}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T13:43:07,500 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-05T13:43:07,501 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733406187474Writing region info on filesystem at 1733406187475 (+1 ms)Initializing all the Stores at 1733406187476 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406187476Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406187477 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406187477Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406187477Cleaning up temporary data from old regions at 1733406187495 (+18 ms)Running coprocessor post-open hooks at 1733406187500 (+5 ms)Region opened successfully at 1733406187501 (+1 ms) 2024-12-05T13:43:07,507 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733406187412 2024-12-05T13:43:07,517 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-05T13:43:07,517 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-05T13:43:07,519 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=da6aa2204f50,34025,1733406185564 2024-12-05T13:43:07,521 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as da6aa2204f50,34025,1733406185564, state=OPEN 2024-12-05T13:43:07,530 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T13:43:07,530 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T13:43:07,530 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T13:43:07,530 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T13:43:07,530 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:07,530 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:07,530 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:07,531 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:07,531 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=da6aa2204f50,34025,1733406185564 2024-12-05T13:43:07,538 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-05T13:43:07,539 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=da6aa2204f50,34025,1733406185564 in 288 msec 2024-12-05T13:43:07,545 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-05T13:43:07,545 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 798 msec 2024-12-05T13:43:07,546 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-05T13:43:07,546 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-05T13:43:07,563 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-05T13:43:07,565 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=da6aa2204f50,34025,1733406185564, seqNum=-1] 2024-12-05T13:43:07,587 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T13:43:07,590 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60297, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T13:43:07,637 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1030 sec 2024-12-05T13:43:07,637 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733406187637, completionTime=-1 2024-12-05T13:43:07,644 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-05T13:43:07,645 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-05T13:43:07,677 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-05T13:43:07,677 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733406247677 2024-12-05T13:43:07,677 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733406307677 2024-12-05T13:43:07,677 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 32 msec 2024-12-05T13:43:07,679 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-05T13:43:07,688 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,36439,1733406184749-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:07,688 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,36439,1733406184749-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:07,689 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,36439,1733406184749-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:07,690 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-da6aa2204f50:36439, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:07,691 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:07,692 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:07,698 DEBUG [master/da6aa2204f50:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-05T13:43:07,721 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.058sec 2024-12-05T13:43:07,723 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-05T13:43:07,725 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-05T13:43:07,726 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-05T13:43:07,727 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-05T13:43:07,727 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-05T13:43:07,728 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,36439,1733406184749-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T13:43:07,728 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,36439,1733406184749-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-05T13:43:07,733 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-05T13:43:07,734 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-05T13:43:07,734 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,36439,1733406184749-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:07,804 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@42b9f9b2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T13:43:07,808 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-05T13:43:07,808 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-05T13:43:07,812 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request da6aa2204f50,36439,-1 for getting cluster id 2024-12-05T13:43:07,815 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-05T13:43:07,824 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'd691aebf-c218-40fc-8a5f-118a6447e57f' 2024-12-05T13:43:07,826 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-05T13:43:07,827 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "d691aebf-c218-40fc-8a5f-118a6447e57f" 2024-12-05T13:43:07,827 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@69c9d640, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T13:43:07,827 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [da6aa2204f50,36439,-1] 2024-12-05T13:43:07,830 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-05T13:43:07,832 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:07,834 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43578, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-05T13:43:07,837 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@523ba0c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T13:43:07,838 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-05T13:43:07,846 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=da6aa2204f50,34025,1733406185564, seqNum=-1] 2024-12-05T13:43:07,847 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T13:43:07,849 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34012, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T13:43:07,868 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=da6aa2204f50,36439,1733406184749 2024-12-05T13:43:07,872 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-05T13:43:07,876 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is da6aa2204f50,36439,1733406184749 2024-12-05T13:43:07,879 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@6a8f1521 2024-12-05T13:43:07,880 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-05T13:43:07,882 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52460, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-05T13:43:07,888 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T13:43:07,897 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-05T13:43:07,900 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T13:43:07,903 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-05T13:43:07,904 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:07,907 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T13:43:07,909 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:07,931 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:07,932 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:07,940 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:58498 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:43151:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:58498 dst: /127.0.0.1:43151 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:07,948 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-05T13:43:07,949 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:07,953 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => bb80ff7c1ab7b36c7d3b13aaeabfd19d, NAME => 'TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130 2024-12-05T13:43:07,965 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:07,966 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:07,973 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:37162 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:37945:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37162 dst: /127.0.0.1:37945 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:07,981 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-05T13:43:07,982 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:07,982 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:07,983 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing bb80ff7c1ab7b36c7d3b13aaeabfd19d, disabling compactions & flushes 2024-12-05T13:43:07,983 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:07,983 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:07,983 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. after waiting 0 ms 2024-12-05T13:43:07,983 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:07,983 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:07,983 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for bb80ff7c1ab7b36c7d3b13aaeabfd19d: Waiting for close lock at 1733406187982Disabling compacts and flushes for region at 1733406187982Disabling writes for close at 1733406187983 (+1 ms)Writing region close event to WAL at 1733406187983Closed at 1733406187983 2024-12-05T13:43:07,987 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T13:43:07,995 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733406187988"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733406187988"}]},"ts":"1733406187988"} 2024-12-05T13:43:08,004 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-05T13:43:08,007 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T13:43:08,011 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733406188007"}]},"ts":"1733406188007"} 2024-12-05T13:43:08,019 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-05T13:43:08,019 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {da6aa2204f50=0} racks are {/default-rack=0} 2024-12-05T13:43:08,021 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-05T13:43:08,021 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-05T13:43:08,021 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-05T13:43:08,021 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-05T13:43:08,022 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-05T13:43:08,022 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-05T13:43:08,022 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-05T13:43:08,022 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-05T13:43:08,022 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-05T13:43:08,022 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-05T13:43:08,024 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:08,024 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bb80ff7c1ab7b36c7d3b13aaeabfd19d, ASSIGN}] 2024-12-05T13:43:08,027 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bb80ff7c1ab7b36c7d3b13aaeabfd19d, ASSIGN 2024-12-05T13:43:08,029 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bb80ff7c1ab7b36c7d3b13aaeabfd19d, ASSIGN; state=OFFLINE, location=da6aa2204f50,39027,1733406185522; forceNewPlan=false, retain=false 2024-12-05T13:43:08,182 INFO [da6aa2204f50:36439 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-05T13:43:08,183 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=bb80ff7c1ab7b36c7d3b13aaeabfd19d, regionState=OPENING, regionLocation=da6aa2204f50,39027,1733406185522 2024-12-05T13:43:08,188 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bb80ff7c1ab7b36c7d3b13aaeabfd19d, ASSIGN because future has completed 2024-12-05T13:43:08,189 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure bb80ff7c1ab7b36c7d3b13aaeabfd19d, server=da6aa2204f50,39027,1733406185522}] 2024-12-05T13:43:08,233 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:08,360 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T13:43:08,365 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:53913, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T13:43:08,376 INFO [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:08,376 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => bb80ff7c1ab7b36c7d3b13aaeabfd19d, NAME => 'TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d.', STARTKEY => '', ENDKEY => ''} 2024-12-05T13:43:08,377 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,377 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:08,377 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,377 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,380 INFO [StoreOpener-bb80ff7c1ab7b36c7d3b13aaeabfd19d-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,383 INFO [StoreOpener-bb80ff7c1ab7b36c7d3b13aaeabfd19d-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region bb80ff7c1ab7b36c7d3b13aaeabfd19d columnFamilyName cf 2024-12-05T13:43:08,383 DEBUG [StoreOpener-bb80ff7c1ab7b36c7d3b13aaeabfd19d-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:08,385 INFO [StoreOpener-bb80ff7c1ab7b36c7d3b13aaeabfd19d-1 {}] regionserver.HStore(327): Store=bb80ff7c1ab7b36c7d3b13aaeabfd19d/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T13:43:08,385 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,387 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,389 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,390 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,390 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,396 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,421 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T13:43:08,422 INFO [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened bb80ff7c1ab7b36c7d3b13aaeabfd19d; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71209172, jitterRate=0.061099350452423096}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T13:43:08,422 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:08,424 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for bb80ff7c1ab7b36c7d3b13aaeabfd19d: Running coprocessor pre-open hook at 1733406188378Writing region info on filesystem at 1733406188378Initializing all the Stores at 1733406188380 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406188380Cleaning up temporary data from old regions at 1733406188390 (+10 ms)Running coprocessor post-open hooks at 1733406188422 (+32 ms)Region opened successfully at 1733406188424 (+2 ms) 2024-12-05T13:43:08,426 INFO [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d., pid=6, masterSystemTime=1733406188359 2024-12-05T13:43:08,431 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:08,432 INFO [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:08,434 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=bb80ff7c1ab7b36c7d3b13aaeabfd19d, regionState=OPEN, openSeqNum=2, regionLocation=da6aa2204f50,39027,1733406185522 2024-12-05T13:43:08,440 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure bb80ff7c1ab7b36c7d3b13aaeabfd19d, server=da6aa2204f50,39027,1733406185522 because future has completed 2024-12-05T13:43:08,458 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-05T13:43:08,459 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure bb80ff7c1ab7b36c7d3b13aaeabfd19d, server=da6aa2204f50,39027,1733406185522 in 257 msec 2024-12-05T13:43:08,467 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-05T13:43:08,467 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bb80ff7c1ab7b36c7d3b13aaeabfd19d, ASSIGN in 435 msec 2024-12-05T13:43:08,471 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T13:43:08,471 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733406188471"}]},"ts":"1733406188471"} 2024-12-05T13:43:08,477 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-05T13:43:08,481 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T13:43:08,491 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 589 msec 2024-12-05T13:43:08,544 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:08,545 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-05T13:43:08,545 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-05T13:43:08,547 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-05T13:43:08,556 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-05T13:43:08,557 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-05T13:43:08,558 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-05T13:43:08,570 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d., hostname=da6aa2204f50,39027,1733406185522, seqNum=2] 2024-12-05T13:43:08,572 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T13:43:08,577 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54194, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T13:43:08,587 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-05T13:43:08,597 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-05T13:43:08,600 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-05T13:43:08,600 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-05T13:43:08,603 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-05T13:43:08,605 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-05T13:43:08,713 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-05T13:43:08,769 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39027 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-05T13:43:08,770 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:08,775 INFO [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing bb80ff7c1ab7b36c7d3b13aaeabfd19d 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-05T13:43:08,888 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d/.tmp/cf/106ae5979c3f4e68914d95c684391b27 is 36, key is row/cf:cq/1733406188578/Put/seqid=0 2024-12-05T13:43:08,900 WARN [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:08,900 WARN [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:08,923 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-05T13:43:08,930 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1194310412_22 at /127.0.0.1:37172 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:37945:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37172 dst: /127.0.0.1:37945 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:08,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-05T13:43:08,944 WARN [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:08,944 INFO [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d/.tmp/cf/106ae5979c3f4e68914d95c684391b27 2024-12-05T13:43:08,999 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d/.tmp/cf/106ae5979c3f4e68914d95c684391b27 as hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d/cf/106ae5979c3f4e68914d95c684391b27 2024-12-05T13:43:09,021 INFO [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d/cf/106ae5979c3f4e68914d95c684391b27, entries=1, sequenceid=5, filesize=4.7 K 2024-12-05T13:43:09,029 INFO [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for bb80ff7c1ab7b36c7d3b13aaeabfd19d in 254ms, sequenceid=5, compaction requested=false 2024-12-05T13:43:09,030 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-05T13:43:09,033 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for bb80ff7c1ab7b36c7d3b13aaeabfd19d: 2024-12-05T13:43:09,033 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:09,035 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-05T13:43:09,038 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-05T13:43:09,052 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-05T13:43:09,052 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 439 msec 2024-12-05T13:43:09,061 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 467 msec 2024-12-05T13:43:09,233 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36439 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-05T13:43:09,233 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-05T13:43:09,252 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-05T13:43:09,253 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-05T13:43:09,253 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:09,259 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,260 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,260 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-05T13:43:09,261 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-05T13:43:09,261 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=437771206, stopped=false 2024-12-05T13:43:09,261 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=da6aa2204f50,36439,1733406184749 2024-12-05T13:43:09,297 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:09,297 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:09,297 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:09,297 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:09,297 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:09,297 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:09,297 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:09,299 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-05T13:43:09,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:09,299 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:09,299 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-05T13:43:09,299 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:09,299 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,299 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'da6aa2204f50,45437,1733406185425' ***** 2024-12-05T13:43:09,300 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-05T13:43:09,300 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'da6aa2204f50,39027,1733406185522' ***** 2024-12-05T13:43:09,300 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-05T13:43:09,300 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'da6aa2204f50,34025,1733406185564' ***** 2024-12-05T13:43:09,300 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-05T13:43:09,301 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T13:43:09,301 INFO [RS:0;da6aa2204f50:45437 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T13:43:09,301 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-05T13:43:09,301 INFO [RS:0;da6aa2204f50:45437 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T13:43:09,301 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(959): stopping server da6aa2204f50,45437,1733406185425 2024-12-05T13:43:09,301 INFO [RS:0;da6aa2204f50:45437 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-05T13:43:09,301 INFO [RS:0;da6aa2204f50:45437 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;da6aa2204f50:45437. 2024-12-05T13:43:09,301 DEBUG [RS:0;da6aa2204f50:45437 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:09,302 DEBUG [RS:0;da6aa2204f50:45437 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,302 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(976): stopping server da6aa2204f50,45437,1733406185425; all regions closed. 2024-12-05T13:43:09,303 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T13:43:09,303 INFO [RS:1;da6aa2204f50:39027 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T13:43:09,303 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-05T13:43:09,303 INFO [RS:1;da6aa2204f50:39027 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T13:43:09,304 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(3091): Received CLOSE for bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:09,304 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:09,304 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T13:43:09,304 INFO [RS:2;da6aa2204f50:34025 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T13:43:09,304 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-05T13:43:09,304 INFO [RS:2;da6aa2204f50:34025 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T13:43:09,304 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(959): stopping server da6aa2204f50,34025,1733406185564 2024-12-05T13:43:09,304 INFO [RS:2;da6aa2204f50:34025 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-05T13:43:09,304 INFO [RS:2;da6aa2204f50:34025 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;da6aa2204f50:34025. 2024-12-05T13:43:09,304 DEBUG [RS:2;da6aa2204f50:34025 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:09,305 DEBUG [RS:2;da6aa2204f50:34025 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,305 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T13:43:09,305 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T13:43:09,305 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T13:43:09,305 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-05T13:43:09,306 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:09,306 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:09,319 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(959): stopping server da6aa2204f50,39027,1733406185522 2024-12-05T13:43:09,319 INFO [RS:1;da6aa2204f50:39027 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-05T13:43:09,319 INFO [RS:1;da6aa2204f50:39027 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;da6aa2204f50:39027. 2024-12-05T13:43:09,319 DEBUG [RS:1;da6aa2204f50:39027 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:09,319 DEBUG [RS:1;da6aa2204f50:39027 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,319 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing bb80ff7c1ab7b36c7d3b13aaeabfd19d, disabling compactions & flushes 2024-12-05T13:43:09,320 INFO [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:09,320 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:09,320 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. after waiting 0 ms 2024-12-05T13:43:09,320 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:09,320 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-05T13:43:09,320 DEBUG [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1325): Online Regions={bb80ff7c1ab7b36c7d3b13aaeabfd19d=TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d.} 2024-12-05T13:43:09,320 DEBUG [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1351): Waiting on bb80ff7c1ab7b36c7d3b13aaeabfd19d 2024-12-05T13:43:09,321 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-05T13:43:09,321 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-05T13:43:09,321 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-05T13:43:09,321 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-05T13:43:09,321 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-05T13:43:09,321 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-05T13:43:09,321 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T13:43:09,321 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T13:43:09,321 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-05T13:43:09,323 INFO [regionserver/da6aa2204f50:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:09,324 INFO [regionserver/da6aa2204f50:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:09,331 INFO [regionserver/da6aa2204f50:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:09,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_1073741828_1018 (size=93) 2024-12-05T13:43:09,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_1073741828_1018 (size=93) 2024-12-05T13:43:09,334 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_1073741828_1018 (size=93) 2024-12-05T13:43:09,344 DEBUG [RS:0;da6aa2204f50:45437 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/oldWALs 2024-12-05T13:43:09,344 INFO [RS:0;da6aa2204f50:45437 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL da6aa2204f50%2C45437%2C1733406185425:(num 1733406186977) 2024-12-05T13:43:09,344 DEBUG [RS:0;da6aa2204f50:45437 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,344 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:09,345 INFO [RS:0;da6aa2204f50:45437 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-05T13:43:09,345 INFO [RS:0;da6aa2204f50:45437 {}] hbase.ChoreService(370): Chore service for: regionserver/da6aa2204f50:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-05T13:43:09,345 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T13:43:09,345 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T13:43:09,345 INFO [regionserver/da6aa2204f50:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-05T13:43:09,345 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T13:43:09,345 INFO [RS:0;da6aa2204f50:45437 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-05T13:43:09,346 INFO [RS:0;da6aa2204f50:45437 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:45437 2024-12-05T13:43:09,355 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T13:43:09,355 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/da6aa2204f50,45437,1733406185425 2024-12-05T13:43:09,355 INFO [RS:0;da6aa2204f50:45437 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-05T13:43:09,363 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [da6aa2204f50,45437,1733406185425] 2024-12-05T13:43:09,373 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/info/eb807561e435403aaadb4bcec50e5ea9 is 153, key is TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d./info:regioninfo/1733406188433/Put/seqid=0 2024-12-05T13:43:09,375 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/default/TestHBaseWalOnEC/bb80ff7c1ab7b36c7d3b13aaeabfd19d/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-05T13:43:09,381 INFO [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:09,382 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for bb80ff7c1ab7b36c7d3b13aaeabfd19d: Waiting for close lock at 1733406189319Running coprocessor pre-close hooks at 1733406189319Disabling compacts and flushes for region at 1733406189319Disabling writes for close at 1733406189320 (+1 ms)Writing region close event to WAL at 1733406189336 (+16 ms)Running coprocessor post-close hooks at 1733406189377 (+41 ms)Closed at 1733406189381 (+4 ms) 2024-12-05T13:43:09,382 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733406187884.bb80ff7c1ab7b36c7d3b13aaeabfd19d. 2024-12-05T13:43:09,383 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:09,383 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:09,396 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_23213930_22 at /127.0.0.1:54638 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:46105:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54638 dst: /127.0.0.1:46105 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:09,405 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-05T13:43:09,408 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:09,408 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/info/eb807561e435403aaadb4bcec50e5ea9 2024-12-05T13:43:09,422 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/da6aa2204f50,45437,1733406185425 already deleted, retry=false 2024-12-05T13:43:09,422 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; da6aa2204f50,45437,1733406185425 expired; onlineServers=2 2024-12-05T13:43:09,454 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/ns/b40e9a7bdd534b28a71c13a8a5c09b88 is 43, key is default/ns:d/1733406187614/Put/seqid=0 2024-12-05T13:43:09,457 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:09,457 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:09,469 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_23213930_22 at /127.0.0.1:54672 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:46105:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54672 dst: /127.0.0.1:46105 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:09,474 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-05T13:43:09,480 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:09,480 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/ns/b40e9a7bdd534b28a71c13a8a5c09b88 2024-12-05T13:43:09,511 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:09,511 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45437-0x101a704f4a10001, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:09,512 INFO [RS:0;da6aa2204f50:45437 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-05T13:43:09,512 INFO [RS:0;da6aa2204f50:45437 {}] regionserver.HRegionServer(1031): Exiting; stopping=da6aa2204f50,45437,1733406185425; zookeeper connection closed. 2024-12-05T13:43:09,515 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7bda8dec {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7bda8dec 2024-12-05T13:43:09,521 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(976): stopping server da6aa2204f50,39027,1733406185522; all regions closed. 2024-12-05T13:43:09,521 DEBUG [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-05T13:43:09,523 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/table/3db0f5f4365c4a9db50b83d947498124 is 52, key is TestHBaseWalOnEC/table:state/1733406188471/Put/seqid=0 2024-12-05T13:43:09,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_1073741826_1016 (size=1298) 2024-12-05T13:43:09,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_1073741826_1016 (size=1298) 2024-12-05T13:43:09,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_1073741826_1016 (size=1298) 2024-12-05T13:43:09,528 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:09,528 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:09,531 DEBUG [RS:1;da6aa2204f50:39027 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/oldWALs 2024-12-05T13:43:09,531 INFO [RS:1;da6aa2204f50:39027 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL da6aa2204f50%2C39027%2C1733406185522:(num 1733406186977) 2024-12-05T13:43:09,531 DEBUG [RS:1;da6aa2204f50:39027 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,531 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:09,532 INFO [RS:1;da6aa2204f50:39027 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-05T13:43:09,532 INFO [RS:1;da6aa2204f50:39027 {}] hbase.ChoreService(370): Chore service for: regionserver/da6aa2204f50:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-05T13:43:09,532 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T13:43:09,532 INFO [regionserver/da6aa2204f50:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-05T13:43:09,532 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T13:43:09,532 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T13:43:09,532 INFO [RS:1;da6aa2204f50:39027 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-05T13:43:09,532 INFO [RS:1;da6aa2204f50:39027 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39027 2024-12-05T13:43:09,537 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_23213930_22 at /127.0.0.1:58536 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:43151:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:58536 dst: /127.0.0.1:43151 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:09,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/da6aa2204f50,39027,1733406185522 2024-12-05T13:43:09,541 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T13:43:09,541 INFO [RS:1;da6aa2204f50:39027 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-05T13:43:09,543 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-05T13:43:09,544 WARN [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:09,544 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/table/3db0f5f4365c4a9db50b83d947498124 2024-12-05T13:43:09,547 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [da6aa2204f50,39027,1733406185522] 2024-12-05T13:43:09,555 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/da6aa2204f50,39027,1733406185522 already deleted, retry=false 2024-12-05T13:43:09,555 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; da6aa2204f50,39027,1733406185522 expired; onlineServers=1 2024-12-05T13:43:09,557 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/info/eb807561e435403aaadb4bcec50e5ea9 as hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/info/eb807561e435403aaadb4bcec50e5ea9 2024-12-05T13:43:09,572 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/info/eb807561e435403aaadb4bcec50e5ea9, entries=10, sequenceid=11, filesize=6.5 K 2024-12-05T13:43:09,574 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/ns/b40e9a7bdd534b28a71c13a8a5c09b88 as hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/ns/b40e9a7bdd534b28a71c13a8a5c09b88 2024-12-05T13:43:09,587 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/ns/b40e9a7bdd534b28a71c13a8a5c09b88, entries=2, sequenceid=11, filesize=5.0 K 2024-12-05T13:43:09,589 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/.tmp/table/3db0f5f4365c4a9db50b83d947498124 as hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/table/3db0f5f4365c4a9db50b83d947498124 2024-12-05T13:43:09,601 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/table/3db0f5f4365c4a9db50b83d947498124, entries=2, sequenceid=11, filesize=5.1 K 2024-12-05T13:43:09,603 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 282ms, sequenceid=11, compaction requested=false 2024-12-05T13:43:09,603 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-05T13:43:09,628 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-05T13:43:09,629 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-05T13:43:09,629 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-05T13:43:09,629 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733406189321Running coprocessor pre-close hooks at 1733406189321Disabling compacts and flushes for region at 1733406189321Disabling writes for close at 1733406189321Obtaining lock to block concurrent updates at 1733406189322 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1733406189322Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733406189322Flushing stores of hbase:meta,,1.1588230740 at 1733406189324 (+2 ms)Flushing 1588230740/info: creating writer at 1733406189324Flushing 1588230740/info: appending metadata at 1733406189369 (+45 ms)Flushing 1588230740/info: closing flushed file at 1733406189369Flushing 1588230740/ns: creating writer at 1733406189428 (+59 ms)Flushing 1588230740/ns: appending metadata at 1733406189452 (+24 ms)Flushing 1588230740/ns: closing flushed file at 1733406189453 (+1 ms)Flushing 1588230740/table: creating writer at 1733406189490 (+37 ms)Flushing 1588230740/table: appending metadata at 1733406189522 (+32 ms)Flushing 1588230740/table: closing flushed file at 1733406189522Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7d7371c1: reopening flushed file at 1733406189555 (+33 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@52c7c278: reopening flushed file at 1733406189572 (+17 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5c21469d: reopening flushed file at 1733406189587 (+15 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 282ms, sequenceid=11, compaction requested=false at 1733406189603 (+16 ms)Writing region close event to WAL at 1733406189606 (+3 ms)Running coprocessor post-close hooks at 1733406189629 (+23 ms)Closed at 1733406189629 2024-12-05T13:43:09,629 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-05T13:43:09,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:09,647 INFO [RS:1;da6aa2204f50:39027 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-05T13:43:09,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39027-0x101a704f4a10002, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:09,647 INFO [RS:1;da6aa2204f50:39027 {}] regionserver.HRegionServer(1031): Exiting; stopping=da6aa2204f50,39027,1733406185522; zookeeper connection closed. 2024-12-05T13:43:09,647 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5618673 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5618673 2024-12-05T13:43:09,721 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(976): stopping server da6aa2204f50,34025,1733406185564; all regions closed. 2024-12-05T13:43:09,727 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_1073741829_1019 (size=2751) 2024-12-05T13:43:09,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_1073741829_1019 (size=2751) 2024-12-05T13:43:09,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_1073741829_1019 (size=2751) 2024-12-05T13:43:09,732 DEBUG [RS:2;da6aa2204f50:34025 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/oldWALs 2024-12-05T13:43:09,732 INFO [RS:2;da6aa2204f50:34025 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL da6aa2204f50%2C34025%2C1733406185564.meta:.meta(num 1733406187442) 2024-12-05T13:43:09,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775773_1004 (size=42) 2024-12-05T13:43:09,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_-9223372036854775772_1004 (size=42) 2024-12-05T13:43:09,752 WARN [Close-WAL-Writer-0 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(650): complete file /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/WALs/da6aa2204f50,34025,1733406185564/da6aa2204f50%2C34025%2C1733406185564.1733406186977 not finished, retry = 0 2024-12-05T13:43:09,756 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_1073741827_1017 (size=93) 2024-12-05T13:43:09,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_1073741827_1017 (size=93) 2024-12-05T13:43:09,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775756_1006 (size=196) 2024-12-05T13:43:09,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_1073741827_1017 (size=93) 2024-12-05T13:43:09,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775757_1006 (size=196) 2024-12-05T13:43:09,783 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_-9223372036854775741_1008 (size=1189) 2024-12-05T13:43:09,783 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775740_1008 (size=1189) 2024-12-05T13:43:09,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_-9223372036854775709_1013 (size=1321) 2024-12-05T13:43:09,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775708_1013 (size=1321) 2024-12-05T13:43:09,822 INFO [regionserver/da6aa2204f50:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-05T13:43:09,822 INFO [regionserver/da6aa2204f50:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-05T13:43:09,856 DEBUG [RS:2;da6aa2204f50:34025 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/oldWALs 2024-12-05T13:43:09,856 INFO [RS:2;da6aa2204f50:34025 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL da6aa2204f50%2C34025%2C1733406185564:(num 1733406186977) 2024-12-05T13:43:09,856 DEBUG [RS:2;da6aa2204f50:34025 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:09,856 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:09,856 INFO [RS:2;da6aa2204f50:34025 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-05T13:43:09,857 INFO [RS:2;da6aa2204f50:34025 {}] hbase.ChoreService(370): Chore service for: regionserver/da6aa2204f50:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-05T13:43:09,857 INFO [RS:2;da6aa2204f50:34025 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-05T13:43:09,857 INFO [regionserver/da6aa2204f50:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-05T13:43:09,857 INFO [RS:2;da6aa2204f50:34025 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:34025 2024-12-05T13:43:09,872 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T13:43:09,872 INFO [RS:2;da6aa2204f50:34025 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-05T13:43:09,872 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/da6aa2204f50,34025,1733406185564 2024-12-05T13:43:09,945 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [da6aa2204f50,34025,1733406185564] 2024-12-05T13:43:09,955 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/da6aa2204f50,34025,1733406185564 already deleted, retry=false 2024-12-05T13:43:09,955 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; da6aa2204f50,34025,1733406185564 expired; onlineServers=0 2024-12-05T13:43:09,956 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'da6aa2204f50,36439,1733406184749' ***** 2024-12-05T13:43:09,956 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-05T13:43:09,956 INFO [M:0;da6aa2204f50:36439 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-05T13:43:09,956 INFO [M:0;da6aa2204f50:36439 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-05T13:43:09,956 DEBUG [M:0;da6aa2204f50:36439 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-05T13:43:09,956 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-05T13:43:09,956 DEBUG [M:0;da6aa2204f50:36439 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-05T13:43:09,956 DEBUG [master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.large.0-1733406186611 {}] cleaner.HFileCleaner(306): Exit Thread[master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.large.0-1733406186611,5,FailOnTimeoutGroup] 2024-12-05T13:43:09,956 DEBUG [master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.small.0-1733406186614 {}] cleaner.HFileCleaner(306): Exit Thread[master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.small.0-1733406186614,5,FailOnTimeoutGroup] 2024-12-05T13:43:09,957 INFO [M:0;da6aa2204f50:36439 {}] hbase.ChoreService(370): Chore service for: master/da6aa2204f50:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-05T13:43:09,957 INFO [M:0;da6aa2204f50:36439 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-05T13:43:09,957 DEBUG [M:0;da6aa2204f50:36439 {}] master.HMaster(1795): Stopping service threads 2024-12-05T13:43:09,957 INFO [M:0;da6aa2204f50:36439 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-05T13:43:09,957 INFO [M:0;da6aa2204f50:36439 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-05T13:43:09,957 INFO [M:0;da6aa2204f50:36439 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-05T13:43:09,957 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-05T13:43:09,963 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-05T13:43:09,964 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:09,964 DEBUG [M:0;da6aa2204f50:36439 {}] zookeeper.ZKUtil(347): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-05T13:43:09,964 WARN [M:0;da6aa2204f50:36439 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-05T13:43:09,965 INFO [M:0;da6aa2204f50:36439 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/.lastflushedseqids 2024-12-05T13:43:09,977 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:09,977 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:09,979 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:54724 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:46105:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54724 dst: /127.0.0.1:46105 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:09,983 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-05T13:43:09,984 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:09,984 INFO [M:0;da6aa2204f50:36439 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-05T13:43:09,984 INFO [M:0;da6aa2204f50:36439 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-05T13:43:09,985 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T13:43:09,985 INFO [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:09,985 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:09,985 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T13:43:09,985 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:09,985 INFO [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.85 KB heapSize=34.13 KB 2024-12-05T13:43:10,005 DEBUG [M:0;da6aa2204f50:36439 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/1502ddbcb84348b48f5bc25ad37626bd is 82, key is hbase:meta,,1/info:regioninfo/1733406187518/Put/seqid=0 2024-12-05T13:43:10,008 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:10,008 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:10,013 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:58604 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:43151:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:58604 dst: /127.0.0.1:43151 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:10,020 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-05T13:43:10,021 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:10,021 INFO [M:0;da6aa2204f50:36439 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/1502ddbcb84348b48f5bc25ad37626bd 2024-12-05T13:43:10,045 INFO [RS:2;da6aa2204f50:34025 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-05T13:43:10,045 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:10,046 INFO [RS:2;da6aa2204f50:34025 {}] regionserver.HRegionServer(1031): Exiting; stopping=da6aa2204f50,34025,1733406185564; zookeeper connection closed. 2024-12-05T13:43:10,046 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34025-0x101a704f4a10003, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:10,046 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@f4cc03a {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@f4cc03a 2024-12-05T13:43:10,046 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-05T13:43:10,055 DEBUG [M:0;da6aa2204f50:36439 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed6ef9470cd54f3590b65a2f4250424f is 749, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733406188487/Put/seqid=0 2024-12-05T13:43:10,057 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:10,057 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:10,063 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:54744 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:46105:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54744 dst: /127.0.0.1:46105 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:10,068 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_-9223372036854775552_1037 (size=6441) 2024-12-05T13:43:10,072 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:10,072 INFO [M:0;da6aa2204f50:36439 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.17 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed6ef9470cd54f3590b65a2f4250424f 2024-12-05T13:43:10,099 DEBUG [M:0;da6aa2204f50:36439 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/d41f6881a46c467c91ace364e16d014b is 69, key is da6aa2204f50,34025,1733406185564/rs:state/1733406186710/Put/seqid=0 2024-12-05T13:43:10,101 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:10,101 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-05T13:43:10,105 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_565424412_22 at /127.0.0.1:37224 [Receiving block BP-75755572-172.17.0.2-1733406181266:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:37945:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37224 dst: /127.0.0.1:37945 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T13:43:10,109 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-05T13:43:10,111 WARN [M:0;da6aa2204f50:36439 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-05T13:43:10,111 INFO [M:0;da6aa2204f50:36439 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/d41f6881a46c467c91ace364e16d014b 2024-12-05T13:43:10,130 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/1502ddbcb84348b48f5bc25ad37626bd as hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/1502ddbcb84348b48f5bc25ad37626bd 2024-12-05T13:43:10,139 INFO [M:0;da6aa2204f50:36439 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/1502ddbcb84348b48f5bc25ad37626bd, entries=8, sequenceid=72, filesize=5.5 K 2024-12-05T13:43:10,141 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed6ef9470cd54f3590b65a2f4250424f as hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ed6ef9470cd54f3590b65a2f4250424f 2024-12-05T13:43:10,151 INFO [M:0;da6aa2204f50:36439 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ed6ef9470cd54f3590b65a2f4250424f, entries=8, sequenceid=72, filesize=6.3 K 2024-12-05T13:43:10,153 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/d41f6881a46c467c91ace364e16d014b as hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/d41f6881a46c467c91ace364e16d014b 2024-12-05T13:43:10,169 INFO [M:0;da6aa2204f50:36439 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/d41f6881a46c467c91ace364e16d014b, entries=3, sequenceid=72, filesize=5.2 K 2024-12-05T13:43:10,171 INFO [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 185ms, sequenceid=72, compaction requested=false 2024-12-05T13:43:10,173 INFO [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:10,173 DEBUG [M:0;da6aa2204f50:36439 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733406189985Disabling compacts and flushes for region at 1733406189985Disabling writes for close at 1733406189985Obtaining lock to block concurrent updates at 1733406189985Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733406189985Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27492, getHeapSize=34888, getOffHeapSize=0, getCellsCount=85 at 1733406189985Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733406189986 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733406189986Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733406190004 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733406190004Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733406190036 (+32 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733406190054 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733406190054Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733406190081 (+27 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733406190098 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733406190099 (+1 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@71683369: reopening flushed file at 1733406190128 (+29 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3d225965: reopening flushed file at 1733406190140 (+12 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@18e05df0: reopening flushed file at 1733406190151 (+11 ms)Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 185ms, sequenceid=72, compaction requested=false at 1733406190171 (+20 ms)Writing region close event to WAL at 1733406190172 (+1 ms)Closed at 1733406190173 (+1 ms) 2024-12-05T13:43:10,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43151 is added to blk_1073741825_1011 (size=32695) 2024-12-05T13:43:10,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37945 is added to blk_1073741825_1011 (size=32695) 2024-12-05T13:43:10,177 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46105 is added to blk_1073741825_1011 (size=32695) 2024-12-05T13:43:10,177 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-05T13:43:10,178 INFO [M:0;da6aa2204f50:36439 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-05T13:43:10,178 INFO [M:0;da6aa2204f50:36439 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36439 2024-12-05T13:43:10,178 INFO [M:0;da6aa2204f50:36439 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-05T13:43:10,287 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:10,287 INFO [M:0;da6aa2204f50:36439 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-05T13:43:10,287 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36439-0x101a704f4a10000, quorum=127.0.0.1:59506, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:10,330 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3114ae69{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:10,333 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3c70a874{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T13:43:10,333 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T13:43:10,333 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5822645a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T13:43:10,333 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@16cd567f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir/,STOPPED} 2024-12-05T13:43:10,336 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T13:43:10,336 WARN [BP-75755572-172.17.0.2-1733406181266 heartbeating to localhost/127.0.0.1:39225 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T13:43:10,336 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T13:43:10,336 WARN [BP-75755572-172.17.0.2-1733406181266 heartbeating to localhost/127.0.0.1:39225 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-75755572-172.17.0.2-1733406181266 (Datanode Uuid 1b53e8ef-7425-4750-a0a5-25afca90c6e6) service to localhost/127.0.0.1:39225 2024-12-05T13:43:10,337 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data5/current/BP-75755572-172.17.0.2-1733406181266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:10,337 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data6/current/BP-75755572-172.17.0.2-1733406181266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:10,338 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T13:43:10,340 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@353955e9{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:10,340 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11738cd8{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T13:43:10,340 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T13:43:10,340 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@40eb7053{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T13:43:10,340 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@510fec09{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir/,STOPPED} 2024-12-05T13:43:10,341 WARN [BP-75755572-172.17.0.2-1733406181266 heartbeating to localhost/127.0.0.1:39225 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T13:43:10,341 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T13:43:10,341 WARN [BP-75755572-172.17.0.2-1733406181266 heartbeating to localhost/127.0.0.1:39225 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-75755572-172.17.0.2-1733406181266 (Datanode Uuid bb905242-3b76-43eb-9622-6d40a1524fde) service to localhost/127.0.0.1:39225 2024-12-05T13:43:10,342 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T13:43:10,342 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data3/current/BP-75755572-172.17.0.2-1733406181266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:10,342 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data4/current/BP-75755572-172.17.0.2-1733406181266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:10,342 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T13:43:10,344 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1b97a472{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:10,344 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3722a29b{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T13:43:10,344 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T13:43:10,345 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@69893329{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T13:43:10,345 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3a5de9e4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir/,STOPPED} 2024-12-05T13:43:10,346 WARN [BP-75755572-172.17.0.2-1733406181266 heartbeating to localhost/127.0.0.1:39225 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T13:43:10,346 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T13:43:10,346 WARN [BP-75755572-172.17.0.2-1733406181266 heartbeating to localhost/127.0.0.1:39225 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-75755572-172.17.0.2-1733406181266 (Datanode Uuid 3a7446c5-e224-41e9-83cb-7e3e10a1909a) service to localhost/127.0.0.1:39225 2024-12-05T13:43:10,346 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T13:43:10,347 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data1/current/BP-75755572-172.17.0.2-1733406181266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:10,347 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/cluster_55000c51-09cd-8917-f38e-3107ede8daad/data/data2/current/BP-75755572-172.17.0.2-1733406181266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:10,347 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T13:43:10,355 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@62d6efd9{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T13:43:10,356 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@353d35a1{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T13:43:10,356 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T13:43:10,356 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@ce709a8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T13:43:10,356 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@760c69c0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir/,STOPPED} 2024-12-05T13:43:10,367 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-05T13:43:10,404 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-05T13:43:10,412 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=92 (was 162), OpenFileDescriptor=441 (was 391) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=269 (was 263) - SystemLoadAverage LEAK? -, ProcessCount=12 (was 11) - ProcessCount LEAK? -, AvailableMemoryMB=8078 (was 8422) 2024-12-05T13:43:10,426 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=92, OpenFileDescriptor=441, MaxFileDescriptor=1048576, SystemLoadAverage=269, ProcessCount=12, AvailableMemoryMB=8075 2024-12-05T13:43:10,426 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-05T13:43:10,427 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.log.dir so I do NOT create it in target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad 2024-12-05T13:43:10,427 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/cc9c1db9-c7f0-27a3-81ab-9c53d83bbeeb/hadoop.tmp.dir so I do NOT create it in target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad 2024-12-05T13:43:10,427 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752, deleteOnExit=true 2024-12-05T13:43:10,427 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-05T13:43:10,428 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/test.cache.data in system properties and HBase conf 2024-12-05T13:43:10,428 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.tmp.dir in system properties and HBase conf 2024-12-05T13:43:10,428 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir in system properties and HBase conf 2024-12-05T13:43:10,428 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-05T13:43:10,428 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-05T13:43:10,428 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-05T13:43:10,428 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-05T13:43:10,429 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-05T13:43:10,429 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-05T13:43:10,429 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-05T13:43:10,429 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T13:43:10,429 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-05T13:43:10,429 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-05T13:43:10,429 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T13:43:10,429 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T13:43:10,430 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-05T13:43:10,430 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/nfs.dump.dir in system properties and HBase conf 2024-12-05T13:43:10,430 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/java.io.tmpdir in system properties and HBase conf 2024-12-05T13:43:10,430 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T13:43:10,430 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-05T13:43:10,430 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-05T13:43:10,739 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:10,746 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T13:43:10,747 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T13:43:10,747 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T13:43:10,748 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T13:43:10,749 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:10,750 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@59bbe271{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir/,AVAILABLE} 2024-12-05T13:43:10,751 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@77ad49ec{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T13:43:10,851 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@75925886{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/java.io.tmpdir/jetty-localhost-35495-hadoop-hdfs-3_4_1-tests_jar-_-any-14898126712928907333/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T13:43:10,852 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@6f1e7dec{HTTP/1.1, (http/1.1)}{localhost:35495} 2024-12-05T13:43:10,852 INFO [Time-limited test {}] server.Server(415): Started @11238ms 2024-12-05T13:43:11,028 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:11,031 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T13:43:11,031 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T13:43:11,032 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T13:43:11,032 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T13:43:11,034 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@46f2e60d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir/,AVAILABLE} 2024-12-05T13:43:11,034 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@436188c4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T13:43:11,135 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6419fd60{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/java.io.tmpdir/jetty-localhost-33423-hadoop-hdfs-3_4_1-tests_jar-_-any-2064497649788679023/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:11,135 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@356b0e7e{HTTP/1.1, (http/1.1)}{localhost:33423} 2024-12-05T13:43:11,135 INFO [Time-limited test {}] server.Server(415): Started @11521ms 2024-12-05T13:43:11,137 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T13:43:11,169 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:11,172 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T13:43:11,174 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T13:43:11,174 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T13:43:11,174 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T13:43:11,175 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6f5c60f4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir/,AVAILABLE} 2024-12-05T13:43:11,175 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6ba59100{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T13:43:11,276 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3c49e2f{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/java.io.tmpdir/jetty-localhost-44873-hadoop-hdfs-3_4_1-tests_jar-_-any-6300328770563482697/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:11,276 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@74a88c50{HTTP/1.1, (http/1.1)}{localhost:44873} 2024-12-05T13:43:11,276 INFO [Time-limited test {}] server.Server(415): Started @11663ms 2024-12-05T13:43:11,278 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T13:43:11,316 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T13:43:11,321 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T13:43:11,322 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T13:43:11,322 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T13:43:11,322 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T13:43:11,323 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@60d8940e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir/,AVAILABLE} 2024-12-05T13:43:11,323 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@10a2b6be{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T13:43:11,429 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@165796ba{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/java.io.tmpdir/jetty-localhost-39237-hadoop-hdfs-3_4_1-tests_jar-_-any-12203801789553734161/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:11,430 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@176751e3{HTTP/1.1, (http/1.1)}{localhost:39237} 2024-12-05T13:43:11,430 INFO [Time-limited test {}] server.Server(415): Started @11816ms 2024-12-05T13:43:11,431 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T13:43:11,830 WARN [Thread-559 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data1/current/BP-1337066470-172.17.0.2-1733406190457/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:11,830 WARN [Thread-560 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data2/current/BP-1337066470-172.17.0.2-1733406190457/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:11,845 WARN [Thread-502 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T13:43:11,849 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x8e32dd6f1039417e with lease ID 0xe10f6cc6f7aca109: Processing first storage report for DS-67df9ae4-6474-451b-9acb-40c8d1793be5 from datanode DatanodeRegistration(127.0.0.1:44273, datanodeUuid=92cdcb55-07da-432c-addb-31d92aad2a70, infoPort=46351, infoSecurePort=0, ipcPort=33987, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457) 2024-12-05T13:43:11,849 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x8e32dd6f1039417e with lease ID 0xe10f6cc6f7aca109: from storage DS-67df9ae4-6474-451b-9acb-40c8d1793be5 node DatanodeRegistration(127.0.0.1:44273, datanodeUuid=92cdcb55-07da-432c-addb-31d92aad2a70, infoPort=46351, infoSecurePort=0, ipcPort=33987, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:11,849 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x8e32dd6f1039417e with lease ID 0xe10f6cc6f7aca109: Processing first storage report for DS-8d7e057d-9c72-4e76-a90f-76e6b4343adf from datanode DatanodeRegistration(127.0.0.1:44273, datanodeUuid=92cdcb55-07da-432c-addb-31d92aad2a70, infoPort=46351, infoSecurePort=0, ipcPort=33987, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457) 2024-12-05T13:43:11,849 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x8e32dd6f1039417e with lease ID 0xe10f6cc6f7aca109: from storage DS-8d7e057d-9c72-4e76-a90f-76e6b4343adf node DatanodeRegistration(127.0.0.1:44273, datanodeUuid=92cdcb55-07da-432c-addb-31d92aad2a70, infoPort=46351, infoSecurePort=0, ipcPort=33987, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:12,133 WARN [Thread-573 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data3/current/BP-1337066470-172.17.0.2-1733406190457/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:12,133 WARN [Thread-574 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data4/current/BP-1337066470-172.17.0.2-1733406190457/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:12,158 WARN [Thread-525 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T13:43:12,161 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2a018e954a2c9fb3 with lease ID 0xe10f6cc6f7aca10a: Processing first storage report for DS-eafd6d2b-2bf5-4398-8841-5adecbf48468 from datanode DatanodeRegistration(127.0.0.1:45111, datanodeUuid=e1b3847a-52cb-4ad8-969d-74e08929030d, infoPort=35449, infoSecurePort=0, ipcPort=34909, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457) 2024-12-05T13:43:12,161 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2a018e954a2c9fb3 with lease ID 0xe10f6cc6f7aca10a: from storage DS-eafd6d2b-2bf5-4398-8841-5adecbf48468 node DatanodeRegistration(127.0.0.1:45111, datanodeUuid=e1b3847a-52cb-4ad8-969d-74e08929030d, infoPort=35449, infoSecurePort=0, ipcPort=34909, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:12,161 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2a018e954a2c9fb3 with lease ID 0xe10f6cc6f7aca10a: Processing first storage report for DS-0dac64ba-4b5c-4125-9c73-021030fae69b from datanode DatanodeRegistration(127.0.0.1:45111, datanodeUuid=e1b3847a-52cb-4ad8-969d-74e08929030d, infoPort=35449, infoSecurePort=0, ipcPort=34909, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457) 2024-12-05T13:43:12,161 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2a018e954a2c9fb3 with lease ID 0xe10f6cc6f7aca10a: from storage DS-0dac64ba-4b5c-4125-9c73-021030fae69b node DatanodeRegistration(127.0.0.1:45111, datanodeUuid=e1b3847a-52cb-4ad8-969d-74e08929030d, infoPort=35449, infoSecurePort=0, ipcPort=34909, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:12,242 WARN [Thread-584 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data5/current/BP-1337066470-172.17.0.2-1733406190457/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:12,242 WARN [Thread-585 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data6/current/BP-1337066470-172.17.0.2-1733406190457/current, will proceed with Du for space computation calculation, 2024-12-05T13:43:12,263 WARN [Thread-548 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T13:43:12,267 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x9a5ccff70915a45b with lease ID 0xe10f6cc6f7aca10b: Processing first storage report for DS-1f0e0d82-5fdd-4c60-9b4b-1b9fdfb347f4 from datanode DatanodeRegistration(127.0.0.1:39543, datanodeUuid=61d983df-7c1c-470a-9f5e-0cc29c879bd3, infoPort=45985, infoSecurePort=0, ipcPort=45103, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457) 2024-12-05T13:43:12,267 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9a5ccff70915a45b with lease ID 0xe10f6cc6f7aca10b: from storage DS-1f0e0d82-5fdd-4c60-9b4b-1b9fdfb347f4 node DatanodeRegistration(127.0.0.1:39543, datanodeUuid=61d983df-7c1c-470a-9f5e-0cc29c879bd3, infoPort=45985, infoSecurePort=0, ipcPort=45103, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:12,267 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x9a5ccff70915a45b with lease ID 0xe10f6cc6f7aca10b: Processing first storage report for DS-2dc0be96-e07b-4ab6-825a-72265d959512 from datanode DatanodeRegistration(127.0.0.1:39543, datanodeUuid=61d983df-7c1c-470a-9f5e-0cc29c879bd3, infoPort=45985, infoSecurePort=0, ipcPort=45103, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457) 2024-12-05T13:43:12,267 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9a5ccff70915a45b with lease ID 0xe10f6cc6f7aca10b: from storage DS-2dc0be96-e07b-4ab6-825a-72265d959512 node DatanodeRegistration(127.0.0.1:39543, datanodeUuid=61d983df-7c1c-470a-9f5e-0cc29c879bd3, infoPort=45985, infoSecurePort=0, ipcPort=45103, storageInfo=lv=-57;cid=testClusterID;nsid=1631998982;c=1733406190457), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T13:43:12,272 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad 2024-12-05T13:43:12,275 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/zookeeper_0, clientPort=50758, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-05T13:43:12,275 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=50758 2024-12-05T13:43:12,276 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,277 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741825_1001 (size=7) 2024-12-05T13:43:12,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741825_1001 (size=7) 2024-12-05T13:43:12,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741825_1001 (size=7) 2024-12-05T13:43:12,292 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025 with version=8 2024-12-05T13:43:12,292 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:39225/user/jenkins/test-data/21ece7f5-d302-4916-a332-f7a43af9a130/hbase-staging 2024-12-05T13:43:12,296 INFO [Time-limited test {}] client.ConnectionUtils(128): master/da6aa2204f50:0 server-side Connection retries=45 2024-12-05T13:43:12,296 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,296 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,296 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T13:43:12,296 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,296 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T13:43:12,296 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-05T13:43:12,297 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T13:43:12,297 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39319 2024-12-05T13:43:12,299 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:39319 connecting to ZooKeeper ensemble=127.0.0.1:50758 2024-12-05T13:43:12,348 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:393190x0, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T13:43:12,349 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:39319-0x101a70515070000 connected 2024-12-05T13:43:12,414 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,416 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,418 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:12,418 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025, hbase.cluster.distributed=false 2024-12-05T13:43:12,420 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T13:43:12,420 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39319 2024-12-05T13:43:12,420 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39319 2024-12-05T13:43:12,421 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39319 2024-12-05T13:43:12,421 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39319 2024-12-05T13:43:12,421 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39319 2024-12-05T13:43:12,436 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/da6aa2204f50:0 server-side Connection retries=45 2024-12-05T13:43:12,436 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,436 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,436 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T13:43:12,436 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,436 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T13:43:12,436 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T13:43:12,436 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T13:43:12,437 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:37501 2024-12-05T13:43:12,439 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:37501 connecting to ZooKeeper ensemble=127.0.0.1:50758 2024-12-05T13:43:12,441 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,444 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,455 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:375010x0, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T13:43:12,456 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:375010x0, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:12,456 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37501-0x101a70515070001 connected 2024-12-05T13:43:12,456 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T13:43:12,457 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T13:43:12,458 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T13:43:12,459 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T13:43:12,459 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37501 2024-12-05T13:43:12,459 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37501 2024-12-05T13:43:12,462 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37501 2024-12-05T13:43:12,463 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37501 2024-12-05T13:43:12,463 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37501 2024-12-05T13:43:12,485 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/da6aa2204f50:0 server-side Connection retries=45 2024-12-05T13:43:12,485 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,485 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,485 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T13:43:12,485 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,485 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T13:43:12,486 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T13:43:12,486 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T13:43:12,487 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:42193 2024-12-05T13:43:12,488 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:42193 connecting to ZooKeeper ensemble=127.0.0.1:50758 2024-12-05T13:43:12,489 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,490 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:421930x0, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T13:43:12,504 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:12,504 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:42193-0x101a70515070002 connected 2024-12-05T13:43:12,504 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T13:43:12,505 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T13:43:12,505 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T13:43:12,506 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T13:43:12,509 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=42193 2024-12-05T13:43:12,509 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=42193 2024-12-05T13:43:12,509 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=42193 2024-12-05T13:43:12,510 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=42193 2024-12-05T13:43:12,510 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=42193 2024-12-05T13:43:12,524 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/da6aa2204f50:0 server-side Connection retries=45 2024-12-05T13:43:12,524 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,524 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,524 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T13:43:12,524 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T13:43:12,524 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T13:43:12,524 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T13:43:12,524 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T13:43:12,525 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:46233 2024-12-05T13:43:12,527 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:46233 connecting to ZooKeeper ensemble=127.0.0.1:50758 2024-12-05T13:43:12,527 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,529 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:462330x0, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T13:43:12,539 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:462330x0, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:12,539 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:46233-0x101a70515070003 connected 2024-12-05T13:43:12,540 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T13:43:12,540 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T13:43:12,541 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T13:43:12,542 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T13:43:12,542 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46233 2024-12-05T13:43:12,542 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46233 2024-12-05T13:43:12,546 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46233 2024-12-05T13:43:12,547 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46233 2024-12-05T13:43:12,547 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46233 2024-12-05T13:43:12,557 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;da6aa2204f50:39319 2024-12-05T13:43:12,558 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/da6aa2204f50,39319,1733406192295 2024-12-05T13:43:12,563 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:12,563 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:12,563 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:12,563 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:12,564 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/da6aa2204f50,39319,1733406192295 2024-12-05T13:43:12,572 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T13:43:12,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T13:43:12,572 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T13:43:12,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,573 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T13:43:12,574 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/da6aa2204f50,39319,1733406192295 from backup master directory 2024-12-05T13:43:12,580 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/da6aa2204f50,39319,1733406192295 2024-12-05T13:43:12,580 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:12,580 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:12,580 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:12,580 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T13:43:12,580 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T13:43:12,580 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=da6aa2204f50,39319,1733406192295 2024-12-05T13:43:12,587 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/hbase.id] with ID: 1c4acf89-568f-4e77-a724-d1b964d7ca25 2024-12-05T13:43:12,587 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/.tmp/hbase.id 2024-12-05T13:43:12,597 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741826_1002 (size=42) 2024-12-05T13:43:12,597 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741826_1002 (size=42) 2024-12-05T13:43:12,598 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741826_1002 (size=42) 2024-12-05T13:43:12,599 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/.tmp/hbase.id]:[hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/hbase.id] 2024-12-05T13:43:12,620 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T13:43:12,620 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-05T13:43:12,622 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-12-05T13:43:12,630 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,630 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,630 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,630 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741827_1003 (size=196) 2024-12-05T13:43:12,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741827_1003 (size=196) 2024-12-05T13:43:12,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741827_1003 (size=196) 2024-12-05T13:43:12,644 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T13:43:12,645 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-05T13:43:12,645 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T13:43:12,659 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741828_1004 (size=1189) 2024-12-05T13:43:12,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741828_1004 (size=1189) 2024-12-05T13:43:12,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741828_1004 (size=1189) 2024-12-05T13:43:12,662 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store 2024-12-05T13:43:12,675 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741829_1005 (size=34) 2024-12-05T13:43:12,675 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741829_1005 (size=34) 2024-12-05T13:43:12,675 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741829_1005 (size=34) 2024-12-05T13:43:12,676 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:12,676 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T13:43:12,676 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:12,676 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:12,676 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T13:43:12,676 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:12,676 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:12,676 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733406192676Disabling compacts and flushes for region at 1733406192676Disabling writes for close at 1733406192676Writing region close event to WAL at 1733406192676Closed at 1733406192676 2024-12-05T13:43:12,677 WARN [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/.initializing 2024-12-05T13:43:12,677 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/WALs/da6aa2204f50,39319,1733406192295 2024-12-05T13:43:12,684 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C39319%2C1733406192295, suffix=, logDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/WALs/da6aa2204f50,39319,1733406192295, archiveDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/oldWALs, maxLogs=10 2024-12-05T13:43:12,685 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor da6aa2204f50%2C39319%2C1733406192295.1733406192684 2024-12-05T13:43:12,698 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/WALs/da6aa2204f50,39319,1733406192295/da6aa2204f50%2C39319%2C1733406192295.1733406192684 2024-12-05T13:43:12,706 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35449:35449),(127.0.0.1/127.0.0.1:45985:45985),(127.0.0.1/127.0.0.1:46351:46351)] 2024-12-05T13:43:12,707 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-05T13:43:12,707 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:12,707 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,707 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,709 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,712 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-05T13:43:12,712 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,713 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:12,713 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,715 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-05T13:43:12,715 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,716 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T13:43:12,716 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,719 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-05T13:43:12,720 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,721 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T13:43:12,721 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,723 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-05T13:43:12,724 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,724 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T13:43:12,725 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,726 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,726 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,728 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,728 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,728 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T13:43:12,730 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T13:43:12,733 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T13:43:12,734 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66523375, jitterRate=-0.00872446596622467}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T13:43:12,735 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733406192707Initializing all the Stores at 1733406192709 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406192709Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406192709Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406192709Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406192709Cleaning up temporary data from old regions at 1733406192728 (+19 ms)Region opened successfully at 1733406192735 (+7 ms) 2024-12-05T13:43:12,735 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-05T13:43:12,742 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4f2fd9e3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=da6aa2204f50/172.17.0.2:0 2024-12-05T13:43:12,743 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-05T13:43:12,744 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-05T13:43:12,744 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-05T13:43:12,744 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-05T13:43:12,745 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-05T13:43:12,746 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-05T13:43:12,746 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-05T13:43:12,749 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-05T13:43:12,750 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-05T13:43:12,755 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-05T13:43:12,756 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-05T13:43:12,757 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-05T13:43:12,763 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-05T13:43:12,764 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-05T13:43:12,766 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-05T13:43:12,772 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-05T13:43:12,773 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-05T13:43:12,780 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-05T13:43:12,783 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-05T13:43:12,788 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-05T13:43:12,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:12,797 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:12,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:12,797 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:12,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,798 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=da6aa2204f50,39319,1733406192295, sessionid=0x101a70515070000, setting cluster-up flag (Was=false) 2024-12-05T13:43:12,814 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,814 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,814 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,814 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,839 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-05T13:43:12,840 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=da6aa2204f50,39319,1733406192295 2024-12-05T13:43:12,855 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,855 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,855 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,855 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:12,880 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-05T13:43:12,881 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=da6aa2204f50,39319,1733406192295 2024-12-05T13:43:12,883 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-05T13:43:12,886 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-05T13:43:12,886 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-05T13:43:12,886 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-05T13:43:12,887 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: da6aa2204f50,39319,1733406192295 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-05T13:43:12,888 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/da6aa2204f50:0, corePoolSize=5, maxPoolSize=5 2024-12-05T13:43:12,888 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/da6aa2204f50:0, corePoolSize=5, maxPoolSize=5 2024-12-05T13:43:12,888 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/da6aa2204f50:0, corePoolSize=5, maxPoolSize=5 2024-12-05T13:43:12,888 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/da6aa2204f50:0, corePoolSize=5, maxPoolSize=5 2024-12-05T13:43:12,888 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/da6aa2204f50:0, corePoolSize=10, maxPoolSize=10 2024-12-05T13:43:12,888 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:12,888 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/da6aa2204f50:0, corePoolSize=2, maxPoolSize=2 2024-12-05T13:43:12,888 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:12,892 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-05T13:43:12,892 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-05T13:43:12,894 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,894 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T13:43:12,897 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733406222897 2024-12-05T13:43:12,898 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-05T13:43:12,898 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-05T13:43:12,898 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-05T13:43:12,898 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-05T13:43:12,898 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-05T13:43:12,898 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-05T13:43:12,900 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:12,902 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-05T13:43:12,902 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-05T13:43:12,902 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-05T13:43:12,906 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-05T13:43:12,906 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-05T13:43:12,906 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.large.0-1733406192906,5,FailOnTimeoutGroup] 2024-12-05T13:43:12,907 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.small.0-1733406192906,5,FailOnTimeoutGroup] 2024-12-05T13:43:12,907 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:12,907 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-05T13:43:12,907 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:12,907 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:12,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741831_1007 (size=1321) 2024-12-05T13:43:12,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741831_1007 (size=1321) 2024-12-05T13:43:12,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741831_1007 (size=1321) 2024-12-05T13:43:12,913 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-05T13:43:12,913 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025 2024-12-05T13:43:12,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741832_1008 (size=32) 2024-12-05T13:43:12,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741832_1008 (size=32) 2024-12-05T13:43:12,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741832_1008 (size=32) 2024-12-05T13:43:12,928 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:12,930 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T13:43:12,932 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T13:43:12,932 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,933 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:12,933 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-05T13:43:12,935 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-05T13:43:12,935 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,936 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:12,936 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T13:43:12,937 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T13:43:12,937 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,938 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:12,938 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T13:43:12,940 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T13:43:12,940 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:12,941 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:12,941 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-05T13:43:12,942 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740 2024-12-05T13:43:12,942 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740 2024-12-05T13:43:12,944 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-05T13:43:12,944 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-05T13:43:12,944 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T13:43:12,945 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-05T13:43:12,948 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T13:43:12,949 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71243279, jitterRate=0.06160758435726166}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T13:43:12,951 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733406192928Initializing all the Stores at 1733406192930 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406192930Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406192930Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406192930Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406192930Cleaning up temporary data from old regions at 1733406192944 (+14 ms)Region opened successfully at 1733406192951 (+7 ms) 2024-12-05T13:43:12,951 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-05T13:43:12,951 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-05T13:43:12,951 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-05T13:43:12,951 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T13:43:12,951 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T13:43:12,957 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(746): ClusterId : 1c4acf89-568f-4e77-a724-d1b964d7ca25 2024-12-05T13:43:12,957 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T13:43:12,957 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-05T13:43:12,957 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733406192951Disabling compacts and flushes for region at 1733406192951Disabling writes for close at 1733406192951Writing region close event to WAL at 1733406192957 (+6 ms)Closed at 1733406192957 2024-12-05T13:43:12,958 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(746): ClusterId : 1c4acf89-568f-4e77-a724-d1b964d7ca25 2024-12-05T13:43:12,958 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(746): ClusterId : 1c4acf89-568f-4e77-a724-d1b964d7ca25 2024-12-05T13:43:12,958 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T13:43:12,958 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T13:43:12,959 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-05T13:43:12,959 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-05T13:43:12,959 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-05T13:43:12,961 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T13:43:12,963 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-05T13:43:12,972 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T13:43:12,973 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T13:43:12,973 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T13:43:12,973 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T13:43:12,973 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T13:43:12,973 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T13:43:12,978 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T13:43:12,981 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T13:43:12,981 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T13:43:12,981 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T13:43:12,981 DEBUG [RS:1;da6aa2204f50:42193 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@641f6d0f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=da6aa2204f50/172.17.0.2:0 2024-12-05T13:43:12,981 DEBUG [RS:2;da6aa2204f50:46233 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@54ca8090, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=da6aa2204f50/172.17.0.2:0 2024-12-05T13:43:12,981 DEBUG [RS:0;da6aa2204f50:37501 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@580c02dd, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=da6aa2204f50/172.17.0.2:0 2024-12-05T13:43:12,984 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T13:43:12,995 DEBUG [RS:2;da6aa2204f50:46233 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;da6aa2204f50:46233 2024-12-05T13:43:12,995 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-05T13:43:12,996 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-05T13:43:12,996 DEBUG [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-05T13:43:12,997 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(2659): reportForDuty to master=da6aa2204f50,39319,1733406192295 with port=46233, startcode=1733406192524 2024-12-05T13:43:12,997 DEBUG [RS:2;da6aa2204f50:46233 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T13:43:13,000 DEBUG [RS:1;da6aa2204f50:42193 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;da6aa2204f50:42193 2024-12-05T13:43:13,001 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-05T13:43:13,001 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-05T13:43:13,001 DEBUG [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-05T13:43:13,001 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(2659): reportForDuty to master=da6aa2204f50,39319,1733406192295 with port=42193, startcode=1733406192484 2024-12-05T13:43:13,002 DEBUG [RS:1;da6aa2204f50:42193 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T13:43:13,003 DEBUG [RS:0;da6aa2204f50:37501 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;da6aa2204f50:37501 2024-12-05T13:43:13,004 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-05T13:43:13,004 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-05T13:43:13,004 DEBUG [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-05T13:43:13,005 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(2659): reportForDuty to master=da6aa2204f50,39319,1733406192295 with port=37501, startcode=1733406192435 2024-12-05T13:43:13,005 DEBUG [RS:0;da6aa2204f50:37501 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T13:43:13,006 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:40727, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T13:43:13,007 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39319 {}] master.ServerManager(363): Checking decommissioned status of RegionServer da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,007 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39319 {}] master.ServerManager(517): Registering regionserver=da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,007 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:39761, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T13:43:13,007 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36299, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T13:43:13,009 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39319 {}] master.ServerManager(363): Checking decommissioned status of RegionServer da6aa2204f50,42193,1733406192484 2024-12-05T13:43:13,009 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39319 {}] master.ServerManager(517): Registering regionserver=da6aa2204f50,42193,1733406192484 2024-12-05T13:43:13,009 DEBUG [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025 2024-12-05T13:43:13,009 DEBUG [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40415 2024-12-05T13:43:13,009 DEBUG [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-05T13:43:13,012 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39319 {}] master.ServerManager(363): Checking decommissioned status of RegionServer da6aa2204f50,37501,1733406192435 2024-12-05T13:43:13,012 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39319 {}] master.ServerManager(517): Registering regionserver=da6aa2204f50,37501,1733406192435 2024-12-05T13:43:13,013 DEBUG [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025 2024-12-05T13:43:13,013 DEBUG [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40415 2024-12-05T13:43:13,013 DEBUG [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-05T13:43:13,014 DEBUG [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025 2024-12-05T13:43:13,014 DEBUG [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40415 2024-12-05T13:43:13,014 DEBUG [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-05T13:43:13,020 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T13:43:13,039 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T13:43:13,040 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T13:43:13,047 DEBUG [RS:2;da6aa2204f50:46233 {}] zookeeper.ZKUtil(111): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,047 WARN [RS:2;da6aa2204f50:46233 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T13:43:13,047 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [da6aa2204f50,37501,1733406192435] 2024-12-05T13:43:13,047 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [da6aa2204f50,46233,1733406192524] 2024-12-05T13:43:13,048 INFO [RS:2;da6aa2204f50:46233 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T13:43:13,048 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [da6aa2204f50,42193,1733406192484] 2024-12-05T13:43:13,048 DEBUG [RS:0;da6aa2204f50:37501 {}] zookeeper.ZKUtil(111): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/da6aa2204f50,37501,1733406192435 2024-12-05T13:43:13,048 DEBUG [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,048 WARN [RS:0;da6aa2204f50:37501 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T13:43:13,048 INFO [RS:0;da6aa2204f50:37501 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T13:43:13,048 DEBUG [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,37501,1733406192435 2024-12-05T13:43:13,049 DEBUG [RS:1;da6aa2204f50:42193 {}] zookeeper.ZKUtil(111): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/da6aa2204f50,42193,1733406192484 2024-12-05T13:43:13,049 WARN [RS:1;da6aa2204f50:42193 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T13:43:13,049 INFO [RS:1;da6aa2204f50:42193 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T13:43:13,049 DEBUG [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,42193,1733406192484 2024-12-05T13:43:13,055 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T13:43:13,055 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T13:43:13,055 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T13:43:13,059 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T13:43:13,061 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T13:43:13,063 INFO [RS:0;da6aa2204f50:37501 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T13:43:13,063 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,065 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T13:43:13,065 INFO [RS:1;da6aa2204f50:42193 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T13:43:13,065 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,066 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-05T13:43:13,066 INFO [RS:2;da6aa2204f50:46233 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T13:43:13,066 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,066 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-05T13:43:13,067 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-05T13:43:13,067 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-05T13:43:13,067 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,067 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,067 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,067 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,067 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,067 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/da6aa2204f50:0, corePoolSize=2, maxPoolSize=2 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:0;da6aa2204f50:37501 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/da6aa2204f50:0, corePoolSize=2, maxPoolSize=2 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,068 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,069 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,069 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:13,069 DEBUG [RS:2;da6aa2204f50:46233 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:13,071 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-05T13:43:13,071 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,071 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,071 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,071 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,071 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/da6aa2204f50:0, corePoolSize=2, maxPoolSize=2 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/da6aa2204f50:0, corePoolSize=1, maxPoolSize=1 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:13,072 DEBUG [RS:1;da6aa2204f50:42193 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0, corePoolSize=3, maxPoolSize=3 2024-12-05T13:43:13,074 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,074 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,074 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,074 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,074 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,074 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,074 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,074 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,37501,1733406192435-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,46233,1733406192524-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,075 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,42193,1733406192484-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T13:43:13,096 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T13:43:13,096 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,46233,1733406192524-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,096 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T13:43:13,096 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,096 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.Replication(171): da6aa2204f50,46233,1733406192524 started 2024-12-05T13:43:13,096 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,37501,1733406192435-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,096 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,097 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.Replication(171): da6aa2204f50,37501,1733406192435 started 2024-12-05T13:43:13,097 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T13:43:13,098 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,42193,1733406192484-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,098 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,098 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.Replication(171): da6aa2204f50,42193,1733406192484 started 2024-12-05T13:43:13,112 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,112 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1482): Serving as da6aa2204f50,46233,1733406192524, RpcServer on da6aa2204f50/172.17.0.2:46233, sessionid=0x101a70515070003 2024-12-05T13:43:13,112 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T13:43:13,112 DEBUG [RS:2;da6aa2204f50:46233 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,112 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,46233,1733406192524' 2024-12-05T13:43:13,112 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T13:43:13,113 WARN [da6aa2204f50:39319 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-05T13:43:13,113 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T13:43:13,114 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T13:43:13,114 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T13:43:13,114 DEBUG [RS:2;da6aa2204f50:46233 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,114 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,46233,1733406192524' 2024-12-05T13:43:13,114 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T13:43:13,115 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T13:43:13,115 DEBUG [RS:2;da6aa2204f50:46233 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T13:43:13,115 INFO [RS:2;da6aa2204f50:46233 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T13:43:13,115 INFO [RS:2;da6aa2204f50:46233 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T13:43:13,115 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,115 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,116 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(1482): Serving as da6aa2204f50,37501,1733406192435, RpcServer on da6aa2204f50/172.17.0.2:37501, sessionid=0x101a70515070001 2024-12-05T13:43:13,116 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(1482): Serving as da6aa2204f50,42193,1733406192484, RpcServer on da6aa2204f50/172.17.0.2:42193, sessionid=0x101a70515070002 2024-12-05T13:43:13,116 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T13:43:13,116 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T13:43:13,116 DEBUG [RS:1;da6aa2204f50:42193 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager da6aa2204f50,42193,1733406192484 2024-12-05T13:43:13,116 DEBUG [RS:0;da6aa2204f50:37501 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager da6aa2204f50,37501,1733406192435 2024-12-05T13:43:13,116 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,42193,1733406192484' 2024-12-05T13:43:13,116 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,37501,1733406192435' 2024-12-05T13:43:13,116 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T13:43:13,116 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T13:43:13,116 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T13:43:13,117 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T13:43:13,117 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T13:43:13,117 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T13:43:13,117 DEBUG [RS:0;da6aa2204f50:37501 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager da6aa2204f50,37501,1733406192435 2024-12-05T13:43:13,117 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T13:43:13,117 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,37501,1733406192435' 2024-12-05T13:43:13,117 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T13:43:13,117 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T13:43:13,117 DEBUG [RS:1;da6aa2204f50:42193 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager da6aa2204f50,42193,1733406192484 2024-12-05T13:43:13,117 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'da6aa2204f50,42193,1733406192484' 2024-12-05T13:43:13,117 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T13:43:13,117 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T13:43:13,118 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T13:43:13,118 DEBUG [RS:0;da6aa2204f50:37501 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T13:43:13,118 INFO [RS:0;da6aa2204f50:37501 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T13:43:13,118 INFO [RS:0;da6aa2204f50:37501 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T13:43:13,118 DEBUG [RS:1;da6aa2204f50:42193 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T13:43:13,118 INFO [RS:1;da6aa2204f50:42193 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T13:43:13,118 INFO [RS:1;da6aa2204f50:42193 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T13:43:13,218 INFO [RS:2;da6aa2204f50:46233 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C46233%2C1733406192524, suffix=, logDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,46233,1733406192524, archiveDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/oldWALs, maxLogs=32 2024-12-05T13:43:13,220 INFO [RS:2;da6aa2204f50:46233 {}] monitor.StreamSlowMonitor(122): New stream slow monitor da6aa2204f50%2C46233%2C1733406192524.1733406193220 2024-12-05T13:43:13,220 INFO [RS:1;da6aa2204f50:42193 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C42193%2C1733406192484, suffix=, logDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,42193,1733406192484, archiveDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/oldWALs, maxLogs=32 2024-12-05T13:43:13,220 INFO [RS:0;da6aa2204f50:37501 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C37501%2C1733406192435, suffix=, logDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,37501,1733406192435, archiveDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/oldWALs, maxLogs=32 2024-12-05T13:43:13,222 INFO [RS:1;da6aa2204f50:42193 {}] monitor.StreamSlowMonitor(122): New stream slow monitor da6aa2204f50%2C42193%2C1733406192484.1733406193222 2024-12-05T13:43:13,222 INFO [RS:0;da6aa2204f50:37501 {}] monitor.StreamSlowMonitor(122): New stream slow monitor da6aa2204f50%2C37501%2C1733406192435.1733406193222 2024-12-05T13:43:13,228 INFO [RS:2;da6aa2204f50:46233 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,46233,1733406192524/da6aa2204f50%2C46233%2C1733406192524.1733406193220 2024-12-05T13:43:13,235 INFO [RS:0;da6aa2204f50:37501 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,37501,1733406192435/da6aa2204f50%2C37501%2C1733406192435.1733406193222 2024-12-05T13:43:13,236 INFO [RS:1;da6aa2204f50:42193 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,42193,1733406192484/da6aa2204f50%2C42193%2C1733406192484.1733406193222 2024-12-05T13:43:13,240 DEBUG [RS:2;da6aa2204f50:46233 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35449:35449),(127.0.0.1/127.0.0.1:46351:46351),(127.0.0.1/127.0.0.1:45985:45985)] 2024-12-05T13:43:13,241 DEBUG [RS:0;da6aa2204f50:37501 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35449:35449),(127.0.0.1/127.0.0.1:45985:45985),(127.0.0.1/127.0.0.1:46351:46351)] 2024-12-05T13:43:13,241 DEBUG [RS:1;da6aa2204f50:42193 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35449:35449),(127.0.0.1/127.0.0.1:46351:46351),(127.0.0.1/127.0.0.1:45985:45985)] 2024-12-05T13:43:13,364 DEBUG [da6aa2204f50:39319 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-05T13:43:13,364 DEBUG [da6aa2204f50:39319 {}] balancer.BalancerClusterState(204): Hosts are {da6aa2204f50=0} racks are {/default-rack=0} 2024-12-05T13:43:13,366 DEBUG [da6aa2204f50:39319 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-05T13:43:13,367 DEBUG [da6aa2204f50:39319 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-05T13:43:13,367 DEBUG [da6aa2204f50:39319 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-05T13:43:13,367 DEBUG [da6aa2204f50:39319 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-05T13:43:13,367 DEBUG [da6aa2204f50:39319 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-05T13:43:13,367 DEBUG [da6aa2204f50:39319 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-05T13:43:13,367 INFO [da6aa2204f50:39319 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-05T13:43:13,367 INFO [da6aa2204f50:39319 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-05T13:43:13,367 INFO [da6aa2204f50:39319 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-05T13:43:13,367 DEBUG [da6aa2204f50:39319 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-05T13:43:13,367 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,369 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as da6aa2204f50,46233,1733406192524, state=OPENING 2024-12-05T13:43:13,420 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-05T13:43:13,430 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:13,430 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:13,430 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:13,430 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:13,431 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:13,431 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:13,431 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:13,431 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T13:43:13,431 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=da6aa2204f50,46233,1733406192524}] 2024-12-05T13:43:13,434 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:13,586 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T13:43:13,588 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37247, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T13:43:13,592 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-05T13:43:13,592 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T13:43:13,595 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=da6aa2204f50%2C46233%2C1733406192524.meta, suffix=.meta, logDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,46233,1733406192524, archiveDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/oldWALs, maxLogs=32 2024-12-05T13:43:13,596 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor da6aa2204f50%2C46233%2C1733406192524.meta.1733406193596.meta 2024-12-05T13:43:13,612 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/WALs/da6aa2204f50,46233,1733406192524/da6aa2204f50%2C46233%2C1733406192524.meta.1733406193596.meta 2024-12-05T13:43:13,613 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35449:35449),(127.0.0.1/127.0.0.1:45985:45985),(127.0.0.1/127.0.0.1:46351:46351)] 2024-12-05T13:43:13,615 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-05T13:43:13,615 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-05T13:43:13,615 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-05T13:43:13,615 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-05T13:43:13,616 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-05T13:43:13,616 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:13,616 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-05T13:43:13,616 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-05T13:43:13,618 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T13:43:13,619 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T13:43:13,619 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:13,619 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:13,620 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-05T13:43:13,621 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-05T13:43:13,621 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:13,621 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:13,622 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T13:43:13,623 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T13:43:13,623 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:13,623 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:13,623 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T13:43:13,624 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T13:43:13,624 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:13,625 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T13:43:13,625 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-05T13:43:13,626 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740 2024-12-05T13:43:13,627 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740 2024-12-05T13:43:13,629 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-05T13:43:13,629 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-05T13:43:13,629 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T13:43:13,631 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-05T13:43:13,632 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66362486, jitterRate=-0.011121898889541626}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T13:43:13,632 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-05T13:43:13,633 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733406193616Writing region info on filesystem at 1733406193616Initializing all the Stores at 1733406193617 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406193617Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406193617Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406193617Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733406193617Cleaning up temporary data from old regions at 1733406193629 (+12 ms)Running coprocessor post-open hooks at 1733406193632 (+3 ms)Region opened successfully at 1733406193633 (+1 ms) 2024-12-05T13:43:13,635 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733406193586 2024-12-05T13:43:13,638 DEBUG [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-05T13:43:13,638 INFO [RS_OPEN_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-05T13:43:13,640 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,641 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as da6aa2204f50,46233,1733406192524, state=OPEN 2024-12-05T13:43:13,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T13:43:13,647 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T13:43:13,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T13:43:13,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T13:43:13,647 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=da6aa2204f50,46233,1733406192524 2024-12-05T13:43:13,647 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:13,647 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:13,647 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:13,647 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T13:43:13,652 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-05T13:43:13,652 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=da6aa2204f50,46233,1733406192524 in 216 msec 2024-12-05T13:43:13,656 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-05T13:43:13,656 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 693 msec 2024-12-05T13:43:13,658 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-05T13:43:13,658 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-05T13:43:13,659 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-05T13:43:13,659 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=da6aa2204f50,46233,1733406192524, seqNum=-1] 2024-12-05T13:43:13,660 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T13:43:13,661 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56977, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T13:43:13,669 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 784 msec 2024-12-05T13:43:13,670 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733406193670, completionTime=-1 2024-12-05T13:43:13,670 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-05T13:43:13,670 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-05T13:43:13,672 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-05T13:43:13,672 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733406253672 2024-12-05T13:43:13,672 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733406313672 2024-12-05T13:43:13,672 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-12-05T13:43:13,672 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-05T13:43:13,673 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,39319,1733406192295-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,673 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,39319,1733406192295-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,673 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,39319,1733406192295-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,673 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-da6aa2204f50:39319, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,673 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,673 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,676 DEBUG [master/da6aa2204f50:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-05T13:43:13,678 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.098sec 2024-12-05T13:43:13,679 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-05T13:43:13,679 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-05T13:43:13,679 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-05T13:43:13,679 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-05T13:43:13,679 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-05T13:43:13,679 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,39319,1733406192295-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T13:43:13,679 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,39319,1733406192295-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-05T13:43:13,681 DEBUG [master/da6aa2204f50:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-05T13:43:13,681 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-05T13:43:13,681 INFO [master/da6aa2204f50:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=da6aa2204f50,39319,1733406192295-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T13:43:13,759 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@62359306, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T13:43:13,759 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request da6aa2204f50,39319,-1 for getting cluster id 2024-12-05T13:43:13,760 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-05T13:43:13,761 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '1c4acf89-568f-4e77-a724-d1b964d7ca25' 2024-12-05T13:43:13,761 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-05T13:43:13,761 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "1c4acf89-568f-4e77-a724-d1b964d7ca25" 2024-12-05T13:43:13,762 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@22cd326c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T13:43:13,762 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [da6aa2204f50,39319,-1] 2024-12-05T13:43:13,762 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-05T13:43:13,762 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:13,764 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37840, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-05T13:43:13,765 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@39a560a9, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T13:43:13,765 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-05T13:43:13,766 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=da6aa2204f50,46233,1733406192524, seqNum=-1] 2024-12-05T13:43:13,767 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T13:43:13,769 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43490, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T13:43:13,771 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=da6aa2204f50,39319,1733406192295 2024-12-05T13:43:13,772 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-05T13:43:13,773 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is da6aa2204f50,39319,1733406192295 2024-12-05T13:43:13,773 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@31ee296d 2024-12-05T13:43:13,774 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-05T13:43:13,776 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37844, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-05T13:43:13,777 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T13:43:13,778 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-05T13:43:13,780 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T13:43:13,780 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:13,781 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-05T13:43:13,782 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:13,782 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T13:43:13,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741837_1013 (size=392) 2024-12-05T13:43:13,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741837_1013 (size=392) 2024-12-05T13:43:13,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741837_1013 (size=392) 2024-12-05T13:43:13,892 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:14,102 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:14,204 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => b559b4469f3f597f9c5fcc21a86bc105, NAME => 'TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025 2024-12-05T13:43:14,212 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741838_1014 (size=51) 2024-12-05T13:43:14,212 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741838_1014 (size=51) 2024-12-05T13:43:14,213 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741838_1014 (size=51) 2024-12-05T13:43:14,214 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:14,214 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing b559b4469f3f597f9c5fcc21a86bc105, disabling compactions & flushes 2024-12-05T13:43:14,214 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:14,214 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:14,214 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. after waiting 0 ms 2024-12-05T13:43:14,214 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:14,214 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:14,214 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for b559b4469f3f597f9c5fcc21a86bc105: Waiting for close lock at 1733406194214Disabling compacts and flushes for region at 1733406194214Disabling writes for close at 1733406194214Writing region close event to WAL at 1733406194214Closed at 1733406194214 2024-12-05T13:43:14,216 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T13:43:14,217 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733406194216"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733406194216"}]},"ts":"1733406194216"} 2024-12-05T13:43:14,220 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-05T13:43:14,222 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T13:43:14,222 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733406194222"}]},"ts":"1733406194222"} 2024-12-05T13:43:14,225 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-05T13:43:14,226 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {da6aa2204f50=0} racks are {/default-rack=0} 2024-12-05T13:43:14,227 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-05T13:43:14,227 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-05T13:43:14,227 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-05T13:43:14,227 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-05T13:43:14,227 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-05T13:43:14,227 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-05T13:43:14,227 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-05T13:43:14,227 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-05T13:43:14,227 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-05T13:43:14,227 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-05T13:43:14,227 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=b559b4469f3f597f9c5fcc21a86bc105, ASSIGN}] 2024-12-05T13:43:14,229 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=b559b4469f3f597f9c5fcc21a86bc105, ASSIGN 2024-12-05T13:43:14,231 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=b559b4469f3f597f9c5fcc21a86bc105, ASSIGN; state=OFFLINE, location=da6aa2204f50,46233,1733406192524; forceNewPlan=false, retain=false 2024-12-05T13:43:14,381 INFO [da6aa2204f50:39319 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-05T13:43:14,382 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=b559b4469f3f597f9c5fcc21a86bc105, regionState=OPENING, regionLocation=da6aa2204f50,46233,1733406192524 2024-12-05T13:43:14,386 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=b559b4469f3f597f9c5fcc21a86bc105, ASSIGN because future has completed 2024-12-05T13:43:14,386 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure b559b4469f3f597f9c5fcc21a86bc105, server=da6aa2204f50,46233,1733406192524}] 2024-12-05T13:43:14,413 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:14,545 INFO [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:14,545 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => b559b4469f3f597f9c5fcc21a86bc105, NAME => 'TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105.', STARTKEY => '', ENDKEY => ''} 2024-12-05T13:43:14,545 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,546 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T13:43:14,546 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,546 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,548 INFO [StoreOpener-b559b4469f3f597f9c5fcc21a86bc105-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,549 INFO [StoreOpener-b559b4469f3f597f9c5fcc21a86bc105-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region b559b4469f3f597f9c5fcc21a86bc105 columnFamilyName cf 2024-12-05T13:43:14,550 DEBUG [StoreOpener-b559b4469f3f597f9c5fcc21a86bc105-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T13:43:14,550 INFO [StoreOpener-b559b4469f3f597f9c5fcc21a86bc105-1 {}] regionserver.HStore(327): Store=b559b4469f3f597f9c5fcc21a86bc105/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T13:43:14,550 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,552 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,552 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,553 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,553 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,555 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,558 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T13:43:14,559 INFO [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened b559b4469f3f597f9c5fcc21a86bc105; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74209227, jitterRate=0.10580365359783173}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T13:43:14,559 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:14,560 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for b559b4469f3f597f9c5fcc21a86bc105: Running coprocessor pre-open hook at 1733406194546Writing region info on filesystem at 1733406194546Initializing all the Stores at 1733406194547 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733406194547Cleaning up temporary data from old regions at 1733406194553 (+6 ms)Running coprocessor post-open hooks at 1733406194559 (+6 ms)Region opened successfully at 1733406194560 (+1 ms) 2024-12-05T13:43:14,561 INFO [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105., pid=6, masterSystemTime=1733406194540 2024-12-05T13:43:14,565 DEBUG [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:14,565 INFO [RS_OPEN_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:14,566 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=b559b4469f3f597f9c5fcc21a86bc105, regionState=OPEN, openSeqNum=2, regionLocation=da6aa2204f50,46233,1733406192524 2024-12-05T13:43:14,569 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure b559b4469f3f597f9c5fcc21a86bc105, server=da6aa2204f50,46233,1733406192524 because future has completed 2024-12-05T13:43:14,575 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-05T13:43:14,575 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure b559b4469f3f597f9c5fcc21a86bc105, server=da6aa2204f50,46233,1733406192524 in 185 msec 2024-12-05T13:43:14,586 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-05T13:43:14,586 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=b559b4469f3f597f9c5fcc21a86bc105, ASSIGN in 348 msec 2024-12-05T13:43:14,589 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T13:43:14,589 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733406194589"}]},"ts":"1733406194589"} 2024-12-05T13:43:14,593 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-05T13:43:14,595 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T13:43:14,599 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 818 msec 2024-12-05T13:43:14,923 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-05T13:43:14,924 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-05T13:43:14,924 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-05T13:43:14,924 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-05T13:43:14,928 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-05T13:43:14,928 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-05T13:43:14,928 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-05T13:43:14,933 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105., hostname=da6aa2204f50,46233,1733406192524, seqNum=2] 2024-12-05T13:43:14,937 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-05T13:43:14,939 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-05T13:43:14,941 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-05T13:43:14,941 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-05T13:43:14,943 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-05T13:43:14,943 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-05T13:43:15,053 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-05T13:43:15,085 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-05T13:43:15,085 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-05T13:43:15,087 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-05T13:43:15,087 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-05T13:43:15,088 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-05T13:43:15,088 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-05T13:43:15,088 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestHBaseWalOnEC 2024-12-05T13:43:15,088 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestHBaseWalOnEC Metrics about Tables on a single HBase RegionServer 2024-12-05T13:43:15,099 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=46233 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-05T13:43:15,099 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:15,099 INFO [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing b559b4469f3f597f9c5fcc21a86bc105 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-05T13:43:15,120 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105/.tmp/cf/52c365e2814d41bb96e666851917f18e is 36, key is row/cf:cq/1733406194934/Put/seqid=0 2024-12-05T13:43:15,121 WARN [IPC Server handler 4 on default port 40415 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-05T13:43:15,122 WARN [IPC Server handler 4 on default port 40415 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-05T13:43:15,122 WARN [IPC Server handler 4 on default port 40415 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-05T13:43:15,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741839_1015 (size=4787) 2024-12-05T13:43:15,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741839_1015 (size=4787) 2024-12-05T13:43:15,133 INFO [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105/.tmp/cf/52c365e2814d41bb96e666851917f18e 2024-12-05T13:43:15,142 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105/.tmp/cf/52c365e2814d41bb96e666851917f18e as hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105/cf/52c365e2814d41bb96e666851917f18e 2024-12-05T13:43:15,151 INFO [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105/cf/52c365e2814d41bb96e666851917f18e, entries=1, sequenceid=5, filesize=4.7 K 2024-12-05T13:43:15,152 INFO [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for b559b4469f3f597f9c5fcc21a86bc105 in 53ms, sequenceid=5, compaction requested=false 2024-12-05T13:43:15,152 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for b559b4469f3f597f9c5fcc21a86bc105: 2024-12-05T13:43:15,153 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:15,153 DEBUG [RS_FLUSH_OPERATIONS-regionserver/da6aa2204f50:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-05T13:43:15,153 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-05T13:43:15,159 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-05T13:43:15,159 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 212 msec 2024-12-05T13:43:15,163 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 223 msec 2024-12-05T13:43:15,262 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39319 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-05T13:43:15,263 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-05T13:43:15,268 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-05T13:43:15,268 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-05T13:43:15,268 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:15,268 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,268 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,268 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-05T13:43:15,268 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-05T13:43:15,268 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1207789188, stopped=false 2024-12-05T13:43:15,269 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=da6aa2204f50,39319,1733406192295 2024-12-05T13:43:15,320 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:15,320 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:15,320 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:15,320 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T13:43:15,320 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:15,320 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:15,320 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:15,320 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:15,320 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-05T13:43:15,321 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-05T13:43:15,321 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:15,321 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,321 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:15,321 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'da6aa2204f50,37501,1733406192435' ***** 2024-12-05T13:43:15,321 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-05T13:43:15,321 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'da6aa2204f50,42193,1733406192484' ***** 2024-12-05T13:43:15,322 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-05T13:43:15,322 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'da6aa2204f50,46233,1733406192524' ***** 2024-12-05T13:43:15,322 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T13:43:15,322 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:15,322 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-05T13:43:15,322 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:15,322 INFO [RS:0;da6aa2204f50:37501 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T13:43:15,322 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-05T13:43:15,322 INFO [RS:0;da6aa2204f50:37501 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T13:43:15,322 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(959): stopping server da6aa2204f50,37501,1733406192435 2024-12-05T13:43:15,322 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T13:43:15,322 INFO [RS:0;da6aa2204f50:37501 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-05T13:43:15,322 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T13:43:15,322 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T13:43:15,322 INFO [RS:2;da6aa2204f50:46233 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T13:43:15,322 INFO [RS:1;da6aa2204f50:42193 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T13:43:15,322 INFO [RS:2;da6aa2204f50:46233 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T13:43:15,322 INFO [RS:1;da6aa2204f50:42193 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T13:43:15,322 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-05T13:43:15,323 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(959): stopping server da6aa2204f50,42193,1733406192484 2024-12-05T13:43:15,322 INFO [RS:0;da6aa2204f50:37501 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;da6aa2204f50:37501. 2024-12-05T13:43:15,323 INFO [RS:1;da6aa2204f50:42193 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-05T13:43:15,322 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-05T13:43:15,323 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(3091): Received CLOSE for b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:15,323 DEBUG [RS:0;da6aa2204f50:37501 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:15,323 DEBUG [RS:0;da6aa2204f50:37501 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,323 INFO [RS:1;da6aa2204f50:42193 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;da6aa2204f50:42193. 2024-12-05T13:43:15,323 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(976): stopping server da6aa2204f50,37501,1733406192435; all regions closed. 2024-12-05T13:43:15,323 DEBUG [RS:1;da6aa2204f50:42193 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:15,323 DEBUG [RS:1;da6aa2204f50:42193 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,323 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(959): stopping server da6aa2204f50,46233,1733406192524 2024-12-05T13:43:15,323 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(976): stopping server da6aa2204f50,42193,1733406192484; all regions closed. 2024-12-05T13:43:15,323 INFO [RS:2;da6aa2204f50:46233 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-05T13:43:15,323 INFO [RS:2;da6aa2204f50:46233 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;da6aa2204f50:46233. 2024-12-05T13:43:15,323 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing b559b4469f3f597f9c5fcc21a86bc105, disabling compactions & flushes 2024-12-05T13:43:15,324 DEBUG [RS:2;da6aa2204f50:46233 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-05T13:43:15,324 INFO [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:15,324 DEBUG [RS:2;da6aa2204f50:46233 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,324 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:15,324 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T13:43:15,324 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T13:43:15,324 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. after waiting 0 ms 2024-12-05T13:43:15,324 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T13:43:15,324 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:15,324 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-05T13:43:15,324 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,324 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,324 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,324 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,324 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,324 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,324 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-05T13:43:15,324 DEBUG [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1325): Online Regions={b559b4469f3f597f9c5fcc21a86bc105=TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105., 1588230740=hbase:meta,,1.1588230740} 2024-12-05T13:43:15,324 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,325 DEBUG [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, b559b4469f3f597f9c5fcc21a86bc105 2024-12-05T13:43:15,325 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,325 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-05T13:43:15,325 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,325 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-05T13:43:15,325 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,325 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-05T13:43:15,325 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T13:43:15,325 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T13:43:15,325 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-05T13:43:15,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741835_1011 (size=93) 2024-12-05T13:43:15,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741834_1010 (size=93) 2024-12-05T13:43:15,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741834_1010 (size=93) 2024-12-05T13:43:15,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741834_1010 (size=93) 2024-12-05T13:43:15,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741835_1011 (size=93) 2024-12-05T13:43:15,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741835_1011 (size=93) 2024-12-05T13:43:15,332 DEBUG [RS:0;da6aa2204f50:37501 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/oldWALs 2024-12-05T13:43:15,332 INFO [RS:0;da6aa2204f50:37501 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog da6aa2204f50%2C37501%2C1733406192435:(num 1733406193222) 2024-12-05T13:43:15,332 DEBUG [RS:0;da6aa2204f50:37501 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,332 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:15,332 INFO [RS:0;da6aa2204f50:37501 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-05T13:43:15,332 DEBUG [RS:1;da6aa2204f50:42193 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/oldWALs 2024-12-05T13:43:15,332 INFO [RS:1;da6aa2204f50:42193 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog da6aa2204f50%2C42193%2C1733406192484:(num 1733406193222) 2024-12-05T13:43:15,332 DEBUG [RS:1;da6aa2204f50:42193 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,332 INFO [RS:0;da6aa2204f50:37501 {}] hbase.ChoreService(370): Chore service for: regionserver/da6aa2204f50:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-05T13:43:15,332 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:15,332 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T13:43:15,332 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T13:43:15,332 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T13:43:15,332 INFO [RS:0;da6aa2204f50:37501 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-05T13:43:15,333 INFO [regionserver/da6aa2204f50:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-05T13:43:15,333 INFO [RS:0;da6aa2204f50:37501 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:37501 2024-12-05T13:43:15,333 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/default/TestHBaseWalOnEC/b559b4469f3f597f9c5fcc21a86bc105/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-05T13:43:15,333 INFO [RS:1;da6aa2204f50:42193 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-05T13:43:15,333 INFO [RS:1;da6aa2204f50:42193 {}] hbase.ChoreService(370): Chore service for: regionserver/da6aa2204f50:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-05T13:43:15,334 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T13:43:15,334 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T13:43:15,334 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T13:43:15,334 INFO [RS:1;da6aa2204f50:42193 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-05T13:43:15,334 INFO [RS:1;da6aa2204f50:42193 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:42193 2024-12-05T13:43:15,334 INFO [regionserver/da6aa2204f50:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-05T13:43:15,335 INFO [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:15,335 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for b559b4469f3f597f9c5fcc21a86bc105: Waiting for close lock at 1733406195323Running coprocessor pre-close hooks at 1733406195323Disabling compacts and flushes for region at 1733406195323Disabling writes for close at 1733406195324 (+1 ms)Writing region close event to WAL at 1733406195325 (+1 ms)Running coprocessor post-close hooks at 1733406195335 (+10 ms)Closed at 1733406195335 2024-12-05T13:43:15,335 DEBUG [RS_CLOSE_REGION-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105. 2024-12-05T13:43:15,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T13:43:15,339 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/da6aa2204f50,37501,1733406192435 2024-12-05T13:43:15,339 INFO [RS:0;da6aa2204f50:37501 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-05T13:43:15,346 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/info/5f5c3590ea2e4d419bbe3e63c64c39c5 is 153, key is TestHBaseWalOnEC,,1733406193776.b559b4469f3f597f9c5fcc21a86bc105./info:regioninfo/1733406194566/Put/seqid=0 2024-12-05T13:43:15,347 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/da6aa2204f50,42193,1733406192484 2024-12-05T13:43:15,347 INFO [RS:1;da6aa2204f50:42193 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-05T13:43:15,355 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [da6aa2204f50,37501,1733406192435] 2024-12-05T13:43:15,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741840_1016 (size=6637) 2024-12-05T13:43:15,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741840_1016 (size=6637) 2024-12-05T13:43:15,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741840_1016 (size=6637) 2024-12-05T13:43:15,358 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/info/5f5c3590ea2e4d419bbe3e63c64c39c5 2024-12-05T13:43:15,372 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/da6aa2204f50,37501,1733406192435 already deleted, retry=false 2024-12-05T13:43:15,372 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; da6aa2204f50,37501,1733406192435 expired; onlineServers=2 2024-12-05T13:43:15,372 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [da6aa2204f50,42193,1733406192484] 2024-12-05T13:43:15,380 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/da6aa2204f50,42193,1733406192484 already deleted, retry=false 2024-12-05T13:43:15,380 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; da6aa2204f50,42193,1733406192484 expired; onlineServers=1 2024-12-05T13:43:15,380 INFO [regionserver/da6aa2204f50:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:15,380 INFO [regionserver/da6aa2204f50:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:15,381 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/ns/db05067d7b1d44dc89d5d48ed1f30af6 is 43, key is default/ns:d/1733406193662/Put/seqid=0 2024-12-05T13:43:15,381 INFO [regionserver/da6aa2204f50:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:15,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741841_1017 (size=5153) 2024-12-05T13:43:15,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741841_1017 (size=5153) 2024-12-05T13:43:15,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741841_1017 (size=5153) 2024-12-05T13:43:15,393 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/ns/db05067d7b1d44dc89d5d48ed1f30af6 2024-12-05T13:43:15,415 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/table/064e53ee2abf4e56a56175a3e7d33733 is 52, key is TestHBaseWalOnEC/table:state/1733406194589/Put/seqid=0 2024-12-05T13:43:15,423 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741842_1018 (size=5249) 2024-12-05T13:43:15,423 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741842_1018 (size=5249) 2024-12-05T13:43:15,423 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741842_1018 (size=5249) 2024-12-05T13:43:15,424 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/table/064e53ee2abf4e56a56175a3e7d33733 2024-12-05T13:43:15,432 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/info/5f5c3590ea2e4d419bbe3e63c64c39c5 as hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/info/5f5c3590ea2e4d419bbe3e63c64c39c5 2024-12-05T13:43:15,440 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/info/5f5c3590ea2e4d419bbe3e63c64c39c5, entries=10, sequenceid=11, filesize=6.5 K 2024-12-05T13:43:15,441 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/ns/db05067d7b1d44dc89d5d48ed1f30af6 as hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/ns/db05067d7b1d44dc89d5d48ed1f30af6 2024-12-05T13:43:15,448 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/ns/db05067d7b1d44dc89d5d48ed1f30af6, entries=2, sequenceid=11, filesize=5.0 K 2024-12-05T13:43:15,449 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/.tmp/table/064e53ee2abf4e56a56175a3e7d33733 as hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/table/064e53ee2abf4e56a56175a3e7d33733 2024-12-05T13:43:15,455 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:15,455 INFO [RS:0;da6aa2204f50:37501 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-05T13:43:15,456 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37501-0x101a70515070001, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:15,456 INFO [RS:0;da6aa2204f50:37501 {}] regionserver.HRegionServer(1031): Exiting; stopping=da6aa2204f50,37501,1733406192435; zookeeper connection closed. 2024-12-05T13:43:15,456 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@49282cf {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@49282cf 2024-12-05T13:43:15,461 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/table/064e53ee2abf4e56a56175a3e7d33733, entries=2, sequenceid=11, filesize=5.1 K 2024-12-05T13:43:15,463 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 138ms, sequenceid=11, compaction requested=false 2024-12-05T13:43:15,464 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:15,464 INFO [RS:1;da6aa2204f50:42193 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-05T13:43:15,464 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42193-0x101a70515070002, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:15,464 INFO [RS:1;da6aa2204f50:42193 {}] regionserver.HRegionServer(1031): Exiting; stopping=da6aa2204f50,42193,1733406192484; zookeeper connection closed. 2024-12-05T13:43:15,464 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5e809068 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5e809068 2024-12-05T13:43:15,473 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-05T13:43:15,474 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-05T13:43:15,474 INFO [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-05T13:43:15,474 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733406195325Running coprocessor pre-close hooks at 1733406195325Disabling compacts and flushes for region at 1733406195325Disabling writes for close at 1733406195325Obtaining lock to block concurrent updates at 1733406195325Preparing flush snapshotting stores in 1588230740 at 1733406195325Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733406195326 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733406195327 (+1 ms)Flushing 1588230740/info: creating writer at 1733406195327Flushing 1588230740/info: appending metadata at 1733406195346 (+19 ms)Flushing 1588230740/info: closing flushed file at 1733406195346Flushing 1588230740/ns: creating writer at 1733406195365 (+19 ms)Flushing 1588230740/ns: appending metadata at 1733406195381 (+16 ms)Flushing 1588230740/ns: closing flushed file at 1733406195381Flushing 1588230740/table: creating writer at 1733406195401 (+20 ms)Flushing 1588230740/table: appending metadata at 1733406195414 (+13 ms)Flushing 1588230740/table: closing flushed file at 1733406195414Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5fdad62a: reopening flushed file at 1733406195431 (+17 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@33a0b68: reopening flushed file at 1733406195440 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1eb3d18a: reopening flushed file at 1733406195448 (+8 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 138ms, sequenceid=11, compaction requested=false at 1733406195463 (+15 ms)Writing region close event to WAL at 1733406195468 (+5 ms)Running coprocessor post-close hooks at 1733406195474 (+6 ms)Closed at 1733406195474 2024-12-05T13:43:15,474 DEBUG [RS_CLOSE_META-regionserver/da6aa2204f50:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-05T13:43:15,525 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(976): stopping server da6aa2204f50,46233,1733406192524; all regions closed. 2024-12-05T13:43:15,525 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,525 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,526 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,526 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,526 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,528 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741836_1012 (size=2751) 2024-12-05T13:43:15,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741836_1012 (size=2751) 2024-12-05T13:43:15,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741836_1012 (size=2751) 2024-12-05T13:43:15,531 DEBUG [RS:2;da6aa2204f50:46233 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/oldWALs 2024-12-05T13:43:15,531 INFO [RS:2;da6aa2204f50:46233 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog da6aa2204f50%2C46233%2C1733406192524.meta:.meta(num 1733406193596) 2024-12-05T13:43:15,532 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,532 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,532 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,532 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,532 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741833_1009 (size=1298) 2024-12-05T13:43:15,538 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741833_1009 (size=1298) 2024-12-05T13:43:15,539 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741833_1009 (size=1298) 2024-12-05T13:43:15,542 DEBUG [RS:2;da6aa2204f50:46233 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/oldWALs 2024-12-05T13:43:15,542 INFO [RS:2;da6aa2204f50:46233 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog da6aa2204f50%2C46233%2C1733406192524:(num 1733406193220) 2024-12-05T13:43:15,542 DEBUG [RS:2;da6aa2204f50:46233 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T13:43:15,542 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T13:43:15,542 INFO [RS:2;da6aa2204f50:46233 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-05T13:43:15,542 INFO [RS:2;da6aa2204f50:46233 {}] hbase.ChoreService(370): Chore service for: regionserver/da6aa2204f50:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-05T13:43:15,542 INFO [RS:2;da6aa2204f50:46233 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-05T13:43:15,542 INFO [regionserver/da6aa2204f50:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-05T13:43:15,542 INFO [RS:2;da6aa2204f50:46233 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:46233 2024-12-05T13:43:15,553 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T13:43:15,553 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/da6aa2204f50,46233,1733406192524 2024-12-05T13:43:15,553 INFO [RS:2;da6aa2204f50:46233 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-05T13:43:15,564 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [da6aa2204f50,46233,1733406192524] 2024-12-05T13:43:15,572 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/da6aa2204f50,46233,1733406192524 already deleted, retry=false 2024-12-05T13:43:15,572 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; da6aa2204f50,46233,1733406192524 expired; onlineServers=0 2024-12-05T13:43:15,572 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'da6aa2204f50,39319,1733406192295' ***** 2024-12-05T13:43:15,572 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-05T13:43:15,572 INFO [M:0;da6aa2204f50:39319 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-05T13:43:15,572 INFO [M:0;da6aa2204f50:39319 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-05T13:43:15,572 DEBUG [M:0;da6aa2204f50:39319 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-05T13:43:15,572 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-05T13:43:15,572 DEBUG [M:0;da6aa2204f50:39319 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-05T13:43:15,572 DEBUG [master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.large.0-1733406192906 {}] cleaner.HFileCleaner(306): Exit Thread[master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.large.0-1733406192906,5,FailOnTimeoutGroup] 2024-12-05T13:43:15,573 DEBUG [master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.small.0-1733406192906 {}] cleaner.HFileCleaner(306): Exit Thread[master/da6aa2204f50:0:becomeActiveMaster-HFileCleaner.small.0-1733406192906,5,FailOnTimeoutGroup] 2024-12-05T13:43:15,573 INFO [M:0;da6aa2204f50:39319 {}] hbase.ChoreService(370): Chore service for: master/da6aa2204f50:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-05T13:43:15,573 INFO [M:0;da6aa2204f50:39319 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-05T13:43:15,573 DEBUG [M:0;da6aa2204f50:39319 {}] master.HMaster(1795): Stopping service threads 2024-12-05T13:43:15,573 INFO [M:0;da6aa2204f50:39319 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-05T13:43:15,573 INFO [M:0;da6aa2204f50:39319 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-05T13:43:15,573 INFO [M:0;da6aa2204f50:39319 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-05T13:43:15,573 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-05T13:43:15,580 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-05T13:43:15,580 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T13:43:15,580 DEBUG [M:0;da6aa2204f50:39319 {}] zookeeper.ZKUtil(347): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-05T13:43:15,580 WARN [M:0;da6aa2204f50:39319 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-05T13:43:15,581 INFO [M:0;da6aa2204f50:39319 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/.lastflushedseqids 2024-12-05T13:43:15,597 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741843_1019 (size=127) 2024-12-05T13:43:15,597 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741843_1019 (size=127) 2024-12-05T13:43:15,597 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741843_1019 (size=127) 2024-12-05T13:43:15,598 INFO [M:0;da6aa2204f50:39319 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-05T13:43:15,598 INFO [M:0;da6aa2204f50:39319 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-05T13:43:15,599 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T13:43:15,599 INFO [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:15,599 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:15,599 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T13:43:15,599 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:15,599 INFO [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.83 KB heapSize=34.12 KB 2024-12-05T13:43:15,616 DEBUG [M:0;da6aa2204f50:39319 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/869ab06597b44d31b5133fbc763b01b5 is 82, key is hbase:meta,,1/info:regioninfo/1733406193639/Put/seqid=0 2024-12-05T13:43:15,626 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741844_1020 (size=5672) 2024-12-05T13:43:15,626 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741844_1020 (size=5672) 2024-12-05T13:43:15,627 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741844_1020 (size=5672) 2024-12-05T13:43:15,628 INFO [M:0;da6aa2204f50:39319 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/869ab06597b44d31b5133fbc763b01b5 2024-12-05T13:43:15,652 DEBUG [M:0;da6aa2204f50:39319 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/3e7a105dc7684a909622461f6f21dc81 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733406194598/Put/seqid=0 2024-12-05T13:43:15,664 INFO [RS:2;da6aa2204f50:46233 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-05T13:43:15,664 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:15,664 INFO [RS:2;da6aa2204f50:46233 {}] regionserver.HRegionServer(1031): Exiting; stopping=da6aa2204f50,46233,1733406192524; zookeeper connection closed. 2024-12-05T13:43:15,664 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46233-0x101a70515070003, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:15,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741845_1021 (size=6439) 2024-12-05T13:43:15,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741845_1021 (size=6439) 2024-12-05T13:43:15,666 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741845_1021 (size=6439) 2024-12-05T13:43:15,666 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@192a2791 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@192a2791 2024-12-05T13:43:15,667 INFO [M:0;da6aa2204f50:39319 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/3e7a105dc7684a909622461f6f21dc81 2024-12-05T13:43:15,667 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-05T13:43:15,689 DEBUG [M:0;da6aa2204f50:39319 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/04a04cc24acb447b9b7f3739d4bb46f2 is 69, key is da6aa2204f50,37501,1733406192435/rs:state/1733406193012/Put/seqid=0 2024-12-05T13:43:15,698 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741846_1022 (size=5294) 2024-12-05T13:43:15,699 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741846_1022 (size=5294) 2024-12-05T13:43:15,699 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741846_1022 (size=5294) 2024-12-05T13:43:15,699 INFO [M:0;da6aa2204f50:39319 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/04a04cc24acb447b9b7f3739d4bb46f2 2024-12-05T13:43:15,709 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/869ab06597b44d31b5133fbc763b01b5 as hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/869ab06597b44d31b5133fbc763b01b5 2024-12-05T13:43:15,717 INFO [M:0;da6aa2204f50:39319 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/869ab06597b44d31b5133fbc763b01b5, entries=8, sequenceid=72, filesize=5.5 K 2024-12-05T13:43:15,718 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/3e7a105dc7684a909622461f6f21dc81 as hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/3e7a105dc7684a909622461f6f21dc81 2024-12-05T13:43:15,724 INFO [M:0;da6aa2204f50:39319 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/3e7a105dc7684a909622461f6f21dc81, entries=8, sequenceid=72, filesize=6.3 K 2024-12-05T13:43:15,725 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/04a04cc24acb447b9b7f3739d4bb46f2 as hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/04a04cc24acb447b9b7f3739d4bb46f2 2024-12-05T13:43:15,732 INFO [M:0;da6aa2204f50:39319 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40415/user/jenkins/test-data/8a2c77c7-493e-12e6-d6d3-e1158e3a9025/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/04a04cc24acb447b9b7f3739d4bb46f2, entries=3, sequenceid=72, filesize=5.2 K 2024-12-05T13:43:15,733 INFO [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 134ms, sequenceid=72, compaction requested=false 2024-12-05T13:43:15,734 INFO [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T13:43:15,734 DEBUG [M:0;da6aa2204f50:39319 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733406195599Disabling compacts and flushes for region at 1733406195599Disabling writes for close at 1733406195599Obtaining lock to block concurrent updates at 1733406195599Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733406195599Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27471, getHeapSize=34872, getOffHeapSize=0, getCellsCount=85 at 1733406195599Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733406195600 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733406195600Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733406195616 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733406195616Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733406195634 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733406195651 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733406195651Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733406195674 (+23 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733406195689 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733406195689Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@25b370a8: reopening flushed file at 1733406195708 (+19 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@40ce010b: reopening flushed file at 1733406195717 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2ae81b8a: reopening flushed file at 1733406195724 (+7 ms)Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 134ms, sequenceid=72, compaction requested=false at 1733406195733 (+9 ms)Writing region close event to WAL at 1733406195734 (+1 ms)Closed at 1733406195734 2024-12-05T13:43:15,735 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,735 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,735 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,735 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,735 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-05T13:43:15,737 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44273 is added to blk_1073741830_1006 (size=32674) 2024-12-05T13:43:15,737 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39543 is added to blk_1073741830_1006 (size=32674) 2024-12-05T13:43:15,738 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45111 is added to blk_1073741830_1006 (size=32674) 2024-12-05T13:43:15,739 INFO [M:0;da6aa2204f50:39319 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-05T13:43:15,739 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-05T13:43:15,739 INFO [M:0;da6aa2204f50:39319 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39319 2024-12-05T13:43:15,739 INFO [M:0;da6aa2204f50:39319 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-05T13:43:15,847 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:15,847 INFO [M:0;da6aa2204f50:39319 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-05T13:43:15,847 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39319-0x101a70515070000, quorum=127.0.0.1:50758, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T13:43:15,849 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@165796ba{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:15,850 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@176751e3{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T13:43:15,850 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T13:43:15,850 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@10a2b6be{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T13:43:15,850 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@60d8940e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir/,STOPPED} 2024-12-05T13:43:15,852 WARN [BP-1337066470-172.17.0.2-1733406190457 heartbeating to localhost/127.0.0.1:40415 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T13:43:15,852 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T13:43:15,852 WARN [BP-1337066470-172.17.0.2-1733406190457 heartbeating to localhost/127.0.0.1:40415 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1337066470-172.17.0.2-1733406190457 (Datanode Uuid 61d983df-7c1c-470a-9f5e-0cc29c879bd3) service to localhost/127.0.0.1:40415 2024-12-05T13:43:15,852 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T13:43:15,853 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data5/current/BP-1337066470-172.17.0.2-1733406190457 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:15,853 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data6/current/BP-1337066470-172.17.0.2-1733406190457 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:15,853 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T13:43:15,855 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3c49e2f{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:15,856 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@74a88c50{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T13:43:15,856 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T13:43:15,856 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6ba59100{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T13:43:15,856 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6f5c60f4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir/,STOPPED} 2024-12-05T13:43:15,857 WARN [BP-1337066470-172.17.0.2-1733406190457 heartbeating to localhost/127.0.0.1:40415 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T13:43:15,857 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T13:43:15,857 WARN [BP-1337066470-172.17.0.2-1733406190457 heartbeating to localhost/127.0.0.1:40415 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1337066470-172.17.0.2-1733406190457 (Datanode Uuid e1b3847a-52cb-4ad8-969d-74e08929030d) service to localhost/127.0.0.1:40415 2024-12-05T13:43:15,857 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T13:43:15,858 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data3/current/BP-1337066470-172.17.0.2-1733406190457 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:15,858 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data4/current/BP-1337066470-172.17.0.2-1733406190457 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:15,859 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T13:43:15,861 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6419fd60{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T13:43:15,861 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@356b0e7e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T13:43:15,861 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T13:43:15,861 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@436188c4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T13:43:15,861 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@46f2e60d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir/,STOPPED} 2024-12-05T13:43:15,863 WARN [BP-1337066470-172.17.0.2-1733406190457 heartbeating to localhost/127.0.0.1:40415 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T13:43:15,863 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T13:43:15,863 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T13:43:15,863 WARN [BP-1337066470-172.17.0.2-1733406190457 heartbeating to localhost/127.0.0.1:40415 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1337066470-172.17.0.2-1733406190457 (Datanode Uuid 92cdcb55-07da-432c-addb-31d92aad2a70) service to localhost/127.0.0.1:40415 2024-12-05T13:43:15,863 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data1/current/BP-1337066470-172.17.0.2-1733406190457 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:15,863 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/cluster_7ef08490-db1c-f067-9f5a-b7a7a84cc752/data/data2/current/BP-1337066470-172.17.0.2-1733406190457 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T13:43:15,864 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T13:43:15,873 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@75925886{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T13:43:15,873 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@6f1e7dec{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T13:43:15,873 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T13:43:15,873 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@77ad49ec{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T13:43:15,873 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@59bbe271{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c58f12d9-1b42-dbfe-513b-ed44b8e0e4ad/hadoop.log.dir/,STOPPED} 2024-12-05T13:43:15,881 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-05T13:43:15,909 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-05T13:43:15,915 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=150 (was 92) - Thread LEAK? -, OpenFileDescriptor=516 (was 441) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=272 (was 269) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 12), AvailableMemoryMB=7827 (was 8075)