2024-12-02 15:44:13,271 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-02 15:44:13,292 main DEBUG Took 0.017206 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-02 15:44:13,292 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-02 15:44:13,293 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-02 15:44:13,295 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-02 15:44:13,296 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,345 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-02 15:44:13,360 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,362 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,363 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,363 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,364 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,364 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,365 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,365 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,366 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,366 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,368 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,369 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,375 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,376 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,376 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,377 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,377 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,378 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,387 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,387 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,388 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,388 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,389 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,390 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 15:44:13,391 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,392 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-02 15:44:13,400 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 15:44:13,402 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-02 15:44:13,404 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-02 15:44:13,404 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-02 15:44:13,406 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-02 15:44:13,407 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-02 15:44:13,424 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-02 15:44:13,428 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-02 15:44:13,430 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-02 15:44:13,433 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-02 15:44:13,434 main DEBUG createAppenders(={Console}) 2024-12-02 15:44:13,435 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-02 15:44:13,435 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-02 15:44:13,439 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-02 15:44:13,440 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-02 15:44:13,440 main DEBUG OutputStream closed 2024-12-02 15:44:13,440 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-02 15:44:13,441 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-02 15:44:13,441 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-02 15:44:13,599 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-02 15:44:13,603 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-02 15:44:13,606 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-02 15:44:13,607 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-02 15:44:13,608 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-02 15:44:13,609 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-02 15:44:13,609 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-02 15:44:13,610 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-02 15:44:13,610 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-02 15:44:13,611 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-02 15:44:13,611 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-02 15:44:13,611 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-02 15:44:13,612 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-02 15:44:13,612 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-02 15:44:13,612 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-02 15:44:13,612 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-02 15:44:13,613 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-02 15:44:13,614 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-02 15:44:13,616 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-02 15:44:13,616 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-02 15:44:13,617 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-02 15:44:13,618 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-02T15:44:13,633 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-02 15:44:13,635 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-02 15:44:13,636 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-02T15:44:13,926 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313 2024-12-02T15:44:13,959 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d, deleteOnExit=true 2024-12-02T15:44:13,961 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/test.cache.data in system properties and HBase conf 2024-12-02T15:44:13,962 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T15:44:13,962 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir in system properties and HBase conf 2024-12-02T15:44:13,963 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T15:44:13,964 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T15:44:13,964 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T15:44:14,057 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-02T15:44:14,165 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T15:44:14,169 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T15:44:14,170 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T15:44:14,170 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T15:44:14,171 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T15:44:14,171 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T15:44:14,172 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T15:44:14,172 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T15:44:14,173 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T15:44:14,174 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T15:44:14,174 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/nfs.dump.dir in system properties and HBase conf 2024-12-02T15:44:14,175 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/java.io.tmpdir in system properties and HBase conf 2024-12-02T15:44:14,175 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T15:44:14,176 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T15:44:14,176 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T15:44:15,179 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-02T15:44:15,261 INFO [Time-limited test {}] log.Log(170): Logging initialized @2922ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-02T15:44:15,336 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:15,412 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T15:44:15,438 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T15:44:15,439 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T15:44:15,440 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T15:44:15,453 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:15,456 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir/,AVAILABLE} 2024-12-02T15:44:15,457 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T15:44:15,638 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@76e4c45c{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/java.io.tmpdir/jetty-localhost-42921-hadoop-hdfs-3_4_1-tests_jar-_-any-5248445184866036449/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T15:44:15,649 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:42921} 2024-12-02T15:44:15,649 INFO [Time-limited test {}] server.Server(415): Started @3311ms 2024-12-02T15:44:16,140 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:16,149 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T15:44:16,150 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T15:44:16,150 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T15:44:16,151 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T15:44:16,151 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@32fec40a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir/,AVAILABLE} 2024-12-02T15:44:16,152 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@17f1c7fc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T15:44:16,251 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7e705dc8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/java.io.tmpdir/jetty-localhost-45397-hadoop-hdfs-3_4_1-tests_jar-_-any-1273045320924544729/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:16,252 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@2ad1569e{HTTP/1.1, (http/1.1)}{localhost:45397} 2024-12-02T15:44:16,252 INFO [Time-limited test {}] server.Server(415): Started @3914ms 2024-12-02T15:44:16,300 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T15:44:16,412 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:16,422 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T15:44:16,424 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T15:44:16,424 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T15:44:16,425 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T15:44:16,426 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4a91ec1e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir/,AVAILABLE} 2024-12-02T15:44:16,427 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2c2c5be{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T15:44:16,532 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@26b068f7{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/java.io.tmpdir/jetty-localhost-38137-hadoop-hdfs-3_4_1-tests_jar-_-any-2935499188309608282/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:16,533 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5739b847{HTTP/1.1, (http/1.1)}{localhost:38137} 2024-12-02T15:44:16,533 INFO [Time-limited test {}] server.Server(415): Started @4195ms 2024-12-02T15:44:16,535 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T15:44:16,567 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:16,572 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T15:44:16,574 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T15:44:16,574 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T15:44:16,574 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T15:44:16,575 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7e1f796{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir/,AVAILABLE} 2024-12-02T15:44:16,576 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1023f385{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T15:44:16,676 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7f750918{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/java.io.tmpdir/jetty-localhost-39517-hadoop-hdfs-3_4_1-tests_jar-_-any-6839016794970959858/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:16,677 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@86bf2a7{HTTP/1.1, (http/1.1)}{localhost:39517} 2024-12-02T15:44:16,677 INFO [Time-limited test {}] server.Server(415): Started @4340ms 2024-12-02T15:44:16,679 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T15:44:17,409 WARN [Thread-125 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data4/current/BP-948202177-172.17.0.3-1733154254750/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:17,409 WARN [Thread-122 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data1/current/BP-948202177-172.17.0.3-1733154254750/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:17,409 WARN [Thread-123 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data2/current/BP-948202177-172.17.0.3-1733154254750/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:17,409 WARN [Thread-124 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data3/current/BP-948202177-172.17.0.3-1733154254750/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:17,426 WARN [Thread-133 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data6/current/BP-948202177-172.17.0.3-1733154254750/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:17,426 WARN [Thread-132 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data5/current/BP-948202177-172.17.0.3-1733154254750/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:17,462 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T15:44:17,463 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T15:44:17,476 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T15:44:17,515 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe3e658ddda84cd58 with lease ID 0x4c750a2417b736bc: Processing first storage report for DS-9fd6a521-3c0d-4bbf-81e7-a76c42cca598 from datanode DatanodeRegistration(127.0.0.1:40719, datanodeUuid=93edeccc-6637-4a92-968d-799c841d9fdb, infoPort=41383, infoSecurePort=0, ipcPort=43991, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750) 2024-12-02T15:44:17,516 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe3e658ddda84cd58 with lease ID 0x4c750a2417b736bc: from storage DS-9fd6a521-3c0d-4bbf-81e7-a76c42cca598 node DatanodeRegistration(127.0.0.1:40719, datanodeUuid=93edeccc-6637-4a92-968d-799c841d9fdb, infoPort=41383, infoSecurePort=0, ipcPort=43991, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T15:44:17,516 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x46b5a3d80c1970fd with lease ID 0x4c750a2417b736bb: Processing first storage report for DS-e3a544b7-2773-454d-9f73-a3c611c95004 from datanode DatanodeRegistration(127.0.0.1:41981, datanodeUuid=98925a14-9d17-469e-9bda-96a2c4e78ae9, infoPort=34621, infoSecurePort=0, ipcPort=41255, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750) 2024-12-02T15:44:17,516 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x46b5a3d80c1970fd with lease ID 0x4c750a2417b736bb: from storage DS-e3a544b7-2773-454d-9f73-a3c611c95004 node DatanodeRegistration(127.0.0.1:41981, datanodeUuid=98925a14-9d17-469e-9bda-96a2c4e78ae9, infoPort=34621, infoSecurePort=0, ipcPort=41255, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:17,516 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x6e1723736851d6b0 with lease ID 0x4c750a2417b736bd: Processing first storage report for DS-2c27ba71-8b49-46d3-8c80-bdf40b0f3bda from datanode DatanodeRegistration(127.0.0.1:42823, datanodeUuid=972a9d42-1c27-4269-a769-8cd739344f28, infoPort=42411, infoSecurePort=0, ipcPort=34879, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750) 2024-12-02T15:44:17,516 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6e1723736851d6b0 with lease ID 0x4c750a2417b736bd: from storage DS-2c27ba71-8b49-46d3-8c80-bdf40b0f3bda node DatanodeRegistration(127.0.0.1:42823, datanodeUuid=972a9d42-1c27-4269-a769-8cd739344f28, infoPort=42411, infoSecurePort=0, ipcPort=34879, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:17,517 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe3e658ddda84cd58 with lease ID 0x4c750a2417b736bc: Processing first storage report for DS-d882ae40-2b23-4fd7-893d-506aa660f819 from datanode DatanodeRegistration(127.0.0.1:40719, datanodeUuid=93edeccc-6637-4a92-968d-799c841d9fdb, infoPort=41383, infoSecurePort=0, ipcPort=43991, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750) 2024-12-02T15:44:17,517 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe3e658ddda84cd58 with lease ID 0x4c750a2417b736bc: from storage DS-d882ae40-2b23-4fd7-893d-506aa660f819 node DatanodeRegistration(127.0.0.1:40719, datanodeUuid=93edeccc-6637-4a92-968d-799c841d9fdb, infoPort=41383, infoSecurePort=0, ipcPort=43991, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:17,517 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x46b5a3d80c1970fd with lease ID 0x4c750a2417b736bb: Processing first storage report for DS-399af6e6-3f20-4662-b6c1-56e4adc146dc from datanode DatanodeRegistration(127.0.0.1:41981, datanodeUuid=98925a14-9d17-469e-9bda-96a2c4e78ae9, infoPort=34621, infoSecurePort=0, ipcPort=41255, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750) 2024-12-02T15:44:17,517 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x46b5a3d80c1970fd with lease ID 0x4c750a2417b736bb: from storage DS-399af6e6-3f20-4662-b6c1-56e4adc146dc node DatanodeRegistration(127.0.0.1:41981, datanodeUuid=98925a14-9d17-469e-9bda-96a2c4e78ae9, infoPort=34621, infoSecurePort=0, ipcPort=41255, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:17,517 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x6e1723736851d6b0 with lease ID 0x4c750a2417b736bd: Processing first storage report for DS-31b44e38-10ff-4d30-9d3f-73e2fbd68ac8 from datanode DatanodeRegistration(127.0.0.1:42823, datanodeUuid=972a9d42-1c27-4269-a769-8cd739344f28, infoPort=42411, infoSecurePort=0, ipcPort=34879, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750) 2024-12-02T15:44:17,517 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6e1723736851d6b0 with lease ID 0x4c750a2417b736bd: from storage DS-31b44e38-10ff-4d30-9d3f-73e2fbd68ac8 node DatanodeRegistration(127.0.0.1:42823, datanodeUuid=972a9d42-1c27-4269-a769-8cd739344f28, infoPort=42411, infoSecurePort=0, ipcPort=34879, storageInfo=lv=-57;cid=testClusterID;nsid=1456618767;c=1733154254750), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T15:44:17,613 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313 2024-12-02T15:44:17,697 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-02T15:44:17,752 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=161, OpenFileDescriptor=393, MaxFileDescriptor=1048576, SystemLoadAverage=652, ProcessCount=11, AvailableMemoryMB=8818 2024-12-02T15:44:17,754 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T15:44:17,761 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-02T15:44:17,841 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/zookeeper_0, clientPort=55083, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T15:44:17,851 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=55083 2024-12-02T15:44:17,874 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:17,878 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:17,989 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:17,990 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:18,036 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:34124 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:40719:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34124 dst: /127.0.0.1:40719 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:18,063 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-02T15:44:18,456 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:18,467 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f with version=8 2024-12-02T15:44:18,468 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/hbase-staging 2024-12-02T15:44:18,546 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-02T15:44:18,799 INFO [Time-limited test {}] client.ConnectionUtils(128): master/e1c0b6c2739c:0 server-side Connection retries=45 2024-12-02T15:44:18,811 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:18,811 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:18,817 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T15:44:18,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:18,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T15:44:18,969 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T15:44:19,031 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-02T15:44:19,039 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-02T15:44:19,043 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T15:44:19,067 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 64946 (auto-detected) 2024-12-02T15:44:19,068 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:03 (auto-detected) 2024-12-02T15:44:19,087 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:43127 2024-12-02T15:44:19,113 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:43127 connecting to ZooKeeper ensemble=127.0.0.1:55083 2024-12-02T15:44:19,224 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:431270x0, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T15:44:19,226 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:43127-0x1019800de120000 connected 2024-12-02T15:44:19,297 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,301 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,313 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:19,318 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f, hbase.cluster.distributed=false 2024-12-02T15:44:19,339 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T15:44:19,343 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43127 2024-12-02T15:44:19,344 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43127 2024-12-02T15:44:19,344 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43127 2024-12-02T15:44:19,345 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43127 2024-12-02T15:44:19,345 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43127 2024-12-02T15:44:19,441 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/e1c0b6c2739c:0 server-side Connection retries=45 2024-12-02T15:44:19,442 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,443 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,443 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T15:44:19,443 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,443 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T15:44:19,446 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T15:44:19,448 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T15:44:19,449 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:45141 2024-12-02T15:44:19,451 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:45141 connecting to ZooKeeper ensemble=127.0.0.1:55083 2024-12-02T15:44:19,452 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,459 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,476 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:451410x0, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T15:44:19,477 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45141-0x1019800de120001 connected 2024-12-02T15:44:19,477 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:19,482 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T15:44:19,491 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T15:44:19,494 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T15:44:19,498 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T15:44:19,499 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45141 2024-12-02T15:44:19,500 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45141 2024-12-02T15:44:19,501 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45141 2024-12-02T15:44:19,502 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45141 2024-12-02T15:44:19,502 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45141 2024-12-02T15:44:19,518 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/e1c0b6c2739c:0 server-side Connection retries=45 2024-12-02T15:44:19,518 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,518 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,519 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T15:44:19,519 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,519 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T15:44:19,519 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T15:44:19,520 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T15:44:19,521 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:34903 2024-12-02T15:44:19,522 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:34903 connecting to ZooKeeper ensemble=127.0.0.1:55083 2024-12-02T15:44:19,523 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,526 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,536 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:349030x0, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T15:44:19,537 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:34903-0x1019800de120002 connected 2024-12-02T15:44:19,537 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:19,537 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T15:44:19,540 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T15:44:19,541 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T15:44:19,544 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T15:44:19,545 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34903 2024-12-02T15:44:19,545 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34903 2024-12-02T15:44:19,546 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34903 2024-12-02T15:44:19,550 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34903 2024-12-02T15:44:19,551 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34903 2024-12-02T15:44:19,569 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/e1c0b6c2739c:0 server-side Connection retries=45 2024-12-02T15:44:19,569 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,569 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,569 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T15:44:19,570 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:19,570 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T15:44:19,570 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T15:44:19,570 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T15:44:19,571 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:38775 2024-12-02T15:44:19,572 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:38775 connecting to ZooKeeper ensemble=127.0.0.1:55083 2024-12-02T15:44:19,574 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,584 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,594 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:387750x0, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T15:44:19,595 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:387750x0, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:19,595 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:38775-0x1019800de120003 connected 2024-12-02T15:44:19,595 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T15:44:19,596 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T15:44:19,597 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T15:44:19,599 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T15:44:19,602 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38775 2024-12-02T15:44:19,602 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38775 2024-12-02T15:44:19,605 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38775 2024-12-02T15:44:19,606 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38775 2024-12-02T15:44:19,608 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38775 2024-12-02T15:44:19,620 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;e1c0b6c2739c:43127 2024-12-02T15:44:19,621 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:19,636 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:19,636 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:19,636 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:19,636 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:19,638 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:19,661 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T15:44:19,661 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T15:44:19,661 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T15:44:19,661 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:19,661 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:19,661 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:19,661 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:19,662 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T15:44:19,663 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/e1c0b6c2739c,43127,1733154258626 from backup master directory 2024-12-02T15:44:19,669 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:19,669 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:19,669 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:19,669 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:19,669 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:19,670 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T15:44:19,670 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:19,672 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-02T15:44:19,674 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-02T15:44:19,739 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/hbase.id] with ID: c866859f-eef4-41fb-926d-a3811ca09816 2024-12-02T15:44:19,739 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/.tmp/hbase.id 2024-12-02T15:44:19,747 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:19,747 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:19,750 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:34152 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:40719:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34152 dst: /127.0.0.1:40719 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:19,760 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-02T15:44:19,761 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:19,762 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/.tmp/hbase.id]:[hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/hbase.id] 2024-12-02T15:44:19,808 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:19,812 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T15:44:19,832 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 17ms. 2024-12-02T15:44:19,844 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:19,844 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:19,844 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:19,844 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:19,862 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:19,862 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:19,869 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:34174 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:40719:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34174 dst: /127.0.0.1:40719 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:19,877 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-02T15:44:19,878 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:19,894 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T15:44:19,897 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T15:44:19,904 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-02T15:44:19,940 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:19,941 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:19,948 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:35524 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:42823:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35524 dst: /127.0.0.1:42823 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:19,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-02T15:44:19,956 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:19,970 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store 2024-12-02T15:44:19,984 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:19,984 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:19,988 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:36514 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:41981:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36514 dst: /127.0.0.1:41981 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:19,994 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-02T15:44:19,995 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:19,998 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-02T15:44:20,001 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:20,002 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T15:44:20,002 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:20,003 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:20,004 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T15:44:20,004 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:20,004 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:20,006 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733154260002Disabling compacts and flushes for region at 1733154260002Disabling writes for close at 1733154260004 (+2 ms)Writing region close event to WAL at 1733154260004Closed at 1733154260004 2024-12-02T15:44:20,008 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/.initializing 2024-12-02T15:44:20,009 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/WALs/e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:20,018 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-02T15:44:20,037 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C43127%2C1733154258626, suffix=, logDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/WALs/e1c0b6c2739c,43127,1733154258626, archiveDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/oldWALs, maxLogs=10 2024-12-02T15:44:20,073 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/WALs/e1c0b6c2739c,43127,1733154258626/e1c0b6c2739c%2C43127%2C1733154258626.1733154260043, exclude list is [], retry=0 2024-12-02T15:44:20,091 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:20,092 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40719,DS-9fd6a521-3c0d-4bbf-81e7-a76c42cca598,DISK] 2024-12-02T15:44:20,092 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42823,DS-2c27ba71-8b49-46d3-8c80-bdf40b0f3bda,DISK] 2024-12-02T15:44:20,092 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41981,DS-e3a544b7-2773-454d-9f73-a3c611c95004,DISK] 2024-12-02T15:44:20,095 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-02T15:44:20,133 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/WALs/e1c0b6c2739c,43127,1733154258626/e1c0b6c2739c%2C43127%2C1733154258626.1733154260043 2024-12-02T15:44:20,134 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34621:34621),(127.0.0.1/127.0.0.1:42411:42411),(127.0.0.1/127.0.0.1:41383:41383)] 2024-12-02T15:44:20,135 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T15:44:20,135 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:20,137 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,138 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,173 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,196 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T15:44:20,200 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,203 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:20,203 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,207 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T15:44:20,208 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,209 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T15:44:20,209 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,212 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T15:44:20,212 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,214 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T15:44:20,214 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,217 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T15:44:20,217 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,219 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T15:44:20,219 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,223 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,225 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,230 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,231 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,234 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T15:44:20,237 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:20,244 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T15:44:20,245 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=62756741, jitterRate=-0.06485168635845184}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T15:44:20,250 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733154260150Initializing all the Stores at 1733154260152 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154260153 (+1 ms)Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154260153Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154260153Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154260153Cleaning up temporary data from old regions at 1733154260231 (+78 ms)Region opened successfully at 1733154260250 (+19 ms) 2024-12-02T15:44:20,252 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T15:44:20,287 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6d84bfe3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=e1c0b6c2739c/172.17.0.3:0 2024-12-02T15:44:20,318 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T15:44:20,328 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T15:44:20,328 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T15:44:20,331 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T15:44:20,332 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-02T15:44:20,337 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 4 msec 2024-12-02T15:44:20,337 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T15:44:20,363 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T15:44:20,373 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T15:44:20,411 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T15:44:20,413 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T15:44:20,414 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T15:44:20,419 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T15:44:20,421 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T15:44:20,425 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T15:44:20,435 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T15:44:20,437 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T15:44:20,444 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T15:44:20,463 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T15:44:20,469 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T15:44:20,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:20,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:20,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:20,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,477 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:20,478 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,478 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,478 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,482 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=e1c0b6c2739c,43127,1733154258626, sessionid=0x1019800de120000, setting cluster-up flag (Was=false) 2024-12-02T15:44:20,510 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,510 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,510 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,511 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-02T15:44:20,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-02T15:44:20,535 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T15:44:20,537 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:20,552 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,552 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,552 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,552 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:20,577 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T15:44:20,579 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:20,587 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T15:44:20,612 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(746): ClusterId : c866859f-eef4-41fb-926d-a3811ca09816 2024-12-02T15:44:20,612 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(746): ClusterId : c866859f-eef4-41fb-926d-a3811ca09816 2024-12-02T15:44:20,614 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T15:44:20,614 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T15:44:20,614 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(746): ClusterId : c866859f-eef4-41fb-926d-a3811ca09816 2024-12-02T15:44:20,614 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T15:44:20,629 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T15:44:20,629 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T15:44:20,629 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T15:44:20,629 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T15:44:20,629 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T15:44:20,629 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T15:44:20,637 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T15:44:20,637 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T15:44:20,637 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T15:44:20,638 DEBUG [RS:2;e1c0b6c2739c:38775 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@42ee9a28, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=e1c0b6c2739c/172.17.0.3:0 2024-12-02T15:44:20,638 DEBUG [RS:1;e1c0b6c2739c:34903 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@260dd99e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=e1c0b6c2739c/172.17.0.3:0 2024-12-02T15:44:20,638 DEBUG [RS:0;e1c0b6c2739c:45141 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7bc7af6d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=e1c0b6c2739c/172.17.0.3:0 2024-12-02T15:44:20,655 DEBUG [RS:0;e1c0b6c2739c:45141 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;e1c0b6c2739c:45141 2024-12-02T15:44:20,657 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;e1c0b6c2739c:34903 2024-12-02T15:44:20,659 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;e1c0b6c2739c:38775 2024-12-02T15:44:20,659 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T15:44:20,659 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T15:44:20,659 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T15:44:20,659 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T15:44:20,659 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T15:44:20,659 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T15:44:20,659 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T15:44:20,659 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T15:44:20,659 DEBUG [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T15:44:20,662 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,43127,1733154258626 with port=38775, startcode=1733154259568 2024-12-02T15:44:20,662 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,43127,1733154258626 with port=45141, startcode=1733154259408 2024-12-02T15:44:20,662 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,43127,1733154258626 with port=34903, startcode=1733154259517 2024-12-02T15:44:20,667 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T15:44:20,676 DEBUG [RS:0;e1c0b6c2739c:45141 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T15:44:20,676 DEBUG [RS:2;e1c0b6c2739c:38775 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T15:44:20,676 DEBUG [RS:1;e1c0b6c2739c:34903 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T15:44:20,679 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T15:44:20,686 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T15:44:20,691 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: e1c0b6c2739c,43127,1733154258626 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T15:44:20,699 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/e1c0b6c2739c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T15:44:20,700 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/e1c0b6c2739c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T15:44:20,700 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T15:44:20,700 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T15:44:20,700 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/e1c0b6c2739c:0, corePoolSize=10, maxPoolSize=10 2024-12-02T15:44:20,700 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:20,700 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T15:44:20,701 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:20,716 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733154290716 2024-12-02T15:44:20,717 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37981, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T15:44:20,717 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41291, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T15:44:20,718 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T15:44:20,720 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59305, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T15:44:20,721 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T15:44:20,722 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T15:44:20,722 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T15:44:20,724 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-02T15:44:20,725 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T15:44:20,726 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T15:44:20,726 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T15:44:20,726 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T15:44:20,728 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,728 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T15:44:20,729 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-02T15:44:20,728 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:20,729 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-02T15:44:20,732 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T15:44:20,734 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T15:44:20,734 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T15:44:20,742 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T15:44:20,742 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T15:44:20,746 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:20,746 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:20,746 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.large.0-1733154260743,5,FailOnTimeoutGroup] 2024-12-02T15:44:20,752 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.small.0-1733154260747,5,FailOnTimeoutGroup] 2024-12-02T15:44:20,752 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:20,752 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T15:44:20,754 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:20,754 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:20,754 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:34216 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:40719:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34216 dst: /127.0.0.1:40719 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:20,756 DEBUG [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-02T15:44:20,756 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-02T15:44:20,756 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-02T15:44:20,756 WARN [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-02T15:44:20,756 WARN [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-02T15:44:20,756 WARN [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-02T15:44:20,763 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-02T15:44:20,764 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:20,766 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T15:44:20,767 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f 2024-12-02T15:44:20,776 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:20,776 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:20,779 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:34234 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:40719:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34234 dst: /127.0.0.1:40719 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:20,792 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-02T15:44:20,794 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:20,795 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:20,799 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T15:44:20,802 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T15:44:20,802 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,804 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:20,804 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T15:44:20,807 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T15:44:20,807 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,809 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:20,809 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T15:44:20,812 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T15:44:20,812 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,814 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:20,814 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T15:44:20,819 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T15:44:20,820 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:20,822 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:20,822 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T15:44:20,824 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740 2024-12-02T15:44:20,825 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740 2024-12-02T15:44:20,831 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T15:44:20,831 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T15:44:20,833 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T15:44:20,838 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T15:44:20,852 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T15:44:20,853 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64617332, jitterRate=-0.03712671995162964}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T15:44:20,856 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733154260796Initializing all the Stores at 1733154260798 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154260798Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154260798Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154260799 (+1 ms)Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154260799Cleaning up temporary data from old regions at 1733154260831 (+32 ms)Region opened successfully at 1733154260856 (+25 ms) 2024-12-02T15:44:20,857 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T15:44:20,857 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T15:44:20,857 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T15:44:20,857 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T15:44:20,857 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T15:44:20,857 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,43127,1733154258626 with port=34903, startcode=1733154259517 2024-12-02T15:44:20,857 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,43127,1733154258626 with port=45141, startcode=1733154259408 2024-12-02T15:44:20,859 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T15:44:20,859 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733154260857Disabling compacts and flushes for region at 1733154260857Disabling writes for close at 1733154260857Writing region close event to WAL at 1733154260858 (+1 ms)Closed at 1733154260859 (+1 ms) 2024-12-02T15:44:20,865 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T15:44:20,865 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T15:44:20,867 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] master.ServerManager(363): Checking decommissioned status of RegionServer e1c0b6c2739c,45141,1733154259408 2024-12-02T15:44:20,869 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,43127,1733154258626 with port=38775, startcode=1733154259568 2024-12-02T15:44:20,869 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] master.ServerManager(517): Registering regionserver=e1c0b6c2739c,45141,1733154259408 2024-12-02T15:44:20,874 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T15:44:20,879 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] master.ServerManager(363): Checking decommissioned status of RegionServer e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:20,880 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] master.ServerManager(517): Registering regionserver=e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:20,880 DEBUG [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f 2024-12-02T15:44:20,880 DEBUG [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:45117 2024-12-02T15:44:20,880 DEBUG [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T15:44:20,885 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] master.ServerManager(363): Checking decommissioned status of RegionServer e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:20,885 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f 2024-12-02T15:44:20,885 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] master.ServerManager(517): Registering regionserver=e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:20,885 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:45117 2024-12-02T15:44:20,885 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T15:44:20,886 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T15:44:20,890 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f 2024-12-02T15:44:20,890 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:45117 2024-12-02T15:44:20,890 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T15:44:20,890 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T15:44:20,892 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T15:44:20,995 DEBUG [RS:0;e1c0b6c2739c:45141 {}] zookeeper.ZKUtil(111): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/e1c0b6c2739c,45141,1733154259408 2024-12-02T15:44:20,995 WARN [RS:0;e1c0b6c2739c:45141 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T15:44:20,995 INFO [RS:0;e1c0b6c2739c:45141 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-02T15:44:20,995 DEBUG [RS:2;e1c0b6c2739c:38775 {}] zookeeper.ZKUtil(111): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:20,995 WARN [RS:2;e1c0b6c2739c:38775 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T15:44:20,995 DEBUG [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,45141,1733154259408 2024-12-02T15:44:20,995 INFO [RS:2;e1c0b6c2739c:38775 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-02T15:44:20,996 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:20,996 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [e1c0b6c2739c,38775,1733154259568] 2024-12-02T15:44:20,997 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [e1c0b6c2739c,34903,1733154259517] 2024-12-02T15:44:20,997 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [e1c0b6c2739c,45141,1733154259408] 2024-12-02T15:44:20,997 DEBUG [RS:1;e1c0b6c2739c:34903 {}] zookeeper.ZKUtil(111): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:20,997 WARN [RS:1;e1c0b6c2739c:34903 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T15:44:20,997 INFO [RS:1;e1c0b6c2739c:34903 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-02T15:44:20,998 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:21,031 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T15:44:21,031 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T15:44:21,031 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T15:44:21,041 WARN [e1c0b6c2739c:43127 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-02T15:44:21,045 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T15:44:21,045 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T15:44:21,048 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T15:44:21,054 INFO [RS:2;e1c0b6c2739c:38775 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T15:44:21,054 INFO [RS:1;e1c0b6c2739c:34903 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T15:44:21,054 INFO [RS:0;e1c0b6c2739c:45141 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T15:44:21,054 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,054 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,054 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,055 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T15:44:21,055 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T15:44:21,056 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T15:44:21,062 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T15:44:21,062 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T15:44:21,062 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T15:44:21,064 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,064 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,064 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,064 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,064 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,064 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,064 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,065 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T15:44:21,065 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,065 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,064 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,066 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,066 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,066 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,066 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,066 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,066 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:21,067 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T15:44:21,067 DEBUG [RS:2;e1c0b6c2739c:38775 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:21,067 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T15:44:21,067 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,067 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,068 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,068 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,068 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,068 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,068 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:21,068 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:21,068 DEBUG [RS:1;e1c0b6c2739c:34903 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:21,068 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:21,068 DEBUG [RS:0;e1c0b6c2739c:45141 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:21,070 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,070 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,38775,1733154259568-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,071 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,072 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,072 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,072 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,072 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,34903,1733154259517-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T15:44:21,072 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,45141,1733154259408-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T15:44:21,091 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T15:44:21,092 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T15:44:21,093 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,38775,1733154259568-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,093 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,34903,1733154259517-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,094 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,094 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,094 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.Replication(171): e1c0b6c2739c,38775,1733154259568 started 2024-12-02T15:44:21,094 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.Replication(171): e1c0b6c2739c,34903,1733154259517 started 2024-12-02T15:44:21,097 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T15:44:21,098 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,45141,1733154259408-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,098 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,098 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.Replication(171): e1c0b6c2739c,45141,1733154259408 started 2024-12-02T15:44:21,111 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,111 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1482): Serving as e1c0b6c2739c,34903,1733154259517, RpcServer on e1c0b6c2739c/172.17.0.3:34903, sessionid=0x1019800de120002 2024-12-02T15:44:21,112 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T15:44:21,112 DEBUG [RS:1;e1c0b6c2739c:34903 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:21,112 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,34903,1733154259517' 2024-12-02T15:44:21,112 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T15:44:21,112 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,112 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1482): Serving as e1c0b6c2739c,38775,1733154259568, RpcServer on e1c0b6c2739c/172.17.0.3:38775, sessionid=0x1019800de120003 2024-12-02T15:44:21,113 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T15:44:21,113 DEBUG [RS:2;e1c0b6c2739c:38775 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:21,113 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,38775,1733154259568' 2024-12-02T15:44:21,113 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T15:44:21,113 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T15:44:21,114 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T15:44:21,114 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T15:44:21,114 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T15:44:21,114 DEBUG [RS:1;e1c0b6c2739c:34903 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:21,114 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,34903,1733154259517' 2024-12-02T15:44:21,114 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T15:44:21,116 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T15:44:21,118 DEBUG [RS:1;e1c0b6c2739c:34903 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T15:44:21,119 INFO [RS:1;e1c0b6c2739c:34903 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T15:44:21,119 INFO [RS:1;e1c0b6c2739c:34903 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T15:44:21,120 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,121 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(1482): Serving as e1c0b6c2739c,45141,1733154259408, RpcServer on e1c0b6c2739c/172.17.0.3:45141, sessionid=0x1019800de120001 2024-12-02T15:44:21,121 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T15:44:21,121 DEBUG [RS:0;e1c0b6c2739c:45141 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager e1c0b6c2739c,45141,1733154259408 2024-12-02T15:44:21,121 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,45141,1733154259408' 2024-12-02T15:44:21,121 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T15:44:21,122 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T15:44:21,123 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T15:44:21,123 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T15:44:21,123 DEBUG [RS:0;e1c0b6c2739c:45141 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager e1c0b6c2739c,45141,1733154259408 2024-12-02T15:44:21,123 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,45141,1733154259408' 2024-12-02T15:44:21,123 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T15:44:21,124 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T15:44:21,126 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T15:44:21,127 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T15:44:21,127 DEBUG [RS:2;e1c0b6c2739c:38775 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:21,127 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,38775,1733154259568' 2024-12-02T15:44:21,127 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T15:44:21,128 DEBUG [RS:0;e1c0b6c2739c:45141 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T15:44:21,128 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T15:44:21,128 INFO [RS:0;e1c0b6c2739c:45141 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T15:44:21,129 INFO [RS:0;e1c0b6c2739c:45141 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T15:44:21,129 DEBUG [RS:2;e1c0b6c2739c:38775 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T15:44:21,129 INFO [RS:2;e1c0b6c2739c:38775 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T15:44:21,129 INFO [RS:2;e1c0b6c2739c:38775 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T15:44:21,224 INFO [RS:1;e1c0b6c2739c:34903 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-02T15:44:21,227 INFO [RS:1;e1c0b6c2739c:34903 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C34903%2C1733154259517, suffix=, logDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,34903,1733154259517, archiveDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/oldWALs, maxLogs=32 2024-12-02T15:44:21,230 INFO [RS:0;e1c0b6c2739c:45141 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-02T15:44:21,233 INFO [RS:2;e1c0b6c2739c:38775 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-02T15:44:21,234 INFO [RS:0;e1c0b6c2739c:45141 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C45141%2C1733154259408, suffix=, logDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,45141,1733154259408, archiveDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/oldWALs, maxLogs=32 2024-12-02T15:44:21,237 INFO [RS:2;e1c0b6c2739c:38775 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C38775%2C1733154259568, suffix=, logDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,38775,1733154259568, archiveDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/oldWALs, maxLogs=32 2024-12-02T15:44:21,248 DEBUG [RS:1;e1c0b6c2739c:34903 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,34903,1733154259517/e1c0b6c2739c%2C34903%2C1733154259517.1733154261229, exclude list is [], retry=0 2024-12-02T15:44:21,254 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40719,DS-9fd6a521-3c0d-4bbf-81e7-a76c42cca598,DISK] 2024-12-02T15:44:21,254 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41981,DS-e3a544b7-2773-454d-9f73-a3c611c95004,DISK] 2024-12-02T15:44:21,255 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42823,DS-2c27ba71-8b49-46d3-8c80-bdf40b0f3bda,DISK] 2024-12-02T15:44:21,255 DEBUG [RS:0;e1c0b6c2739c:45141 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,45141,1733154259408/e1c0b6c2739c%2C45141%2C1733154259408.1733154261236, exclude list is [], retry=0 2024-12-02T15:44:21,267 INFO [RS:1;e1c0b6c2739c:34903 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,34903,1733154259517/e1c0b6c2739c%2C34903%2C1733154259517.1733154261229 2024-12-02T15:44:21,268 DEBUG [RS:1;e1c0b6c2739c:34903 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:41383:41383),(127.0.0.1/127.0.0.1:42411:42411),(127.0.0.1/127.0.0.1:34621:34621)] 2024-12-02T15:44:21,288 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40719,DS-9fd6a521-3c0d-4bbf-81e7-a76c42cca598,DISK] 2024-12-02T15:44:21,289 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42823,DS-2c27ba71-8b49-46d3-8c80-bdf40b0f3bda,DISK] 2024-12-02T15:44:21,289 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41981,DS-e3a544b7-2773-454d-9f73-a3c611c95004,DISK] 2024-12-02T15:44:21,290 DEBUG [RS:2;e1c0b6c2739c:38775 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,38775,1733154259568/e1c0b6c2739c%2C38775%2C1733154259568.1733154261242, exclude list is [], retry=0 2024-12-02T15:44:21,294 DEBUG [e1c0b6c2739c:43127 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T15:44:21,298 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42823,DS-2c27ba71-8b49-46d3-8c80-bdf40b0f3bda,DISK] 2024-12-02T15:44:21,298 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40719,DS-9fd6a521-3c0d-4bbf-81e7-a76c42cca598,DISK] 2024-12-02T15:44:21,298 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41981,DS-e3a544b7-2773-454d-9f73-a3c611c95004,DISK] 2024-12-02T15:44:21,307 INFO [RS:0;e1c0b6c2739c:45141 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,45141,1733154259408/e1c0b6c2739c%2C45141%2C1733154259408.1733154261236 2024-12-02T15:44:21,311 DEBUG [RS:0;e1c0b6c2739c:45141 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34621:34621),(127.0.0.1/127.0.0.1:42411:42411),(127.0.0.1/127.0.0.1:41383:41383)] 2024-12-02T15:44:21,312 INFO [RS:2;e1c0b6c2739c:38775 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,38775,1733154259568/e1c0b6c2739c%2C38775%2C1733154259568.1733154261242 2024-12-02T15:44:21,312 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:21,320 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as e1c0b6c2739c,34903,1733154259517, state=OPENING 2024-12-02T15:44:21,322 DEBUG [RS:2;e1c0b6c2739c:38775 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34621:34621),(127.0.0.1/127.0.0.1:42411:42411),(127.0.0.1/127.0.0.1:41383:41383)] 2024-12-02T15:44:21,344 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T15:44:21,352 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:21,352 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:21,352 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:21,352 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:21,353 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:21,353 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:21,353 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:21,353 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:21,355 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T15:44:21,358 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=e1c0b6c2739c,34903,1733154259517}] 2024-12-02T15:44:21,532 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T15:44:21,535 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45679, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T15:44:21,547 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T15:44:21,548 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-02T15:44:21,548 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-02T15:44:21,552 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C34903%2C1733154259517.meta, suffix=.meta, logDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,34903,1733154259517, archiveDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/oldWALs, maxLogs=32 2024-12-02T15:44:21,574 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,34903,1733154259517/e1c0b6c2739c%2C34903%2C1733154259517.meta.1733154261554.meta, exclude list is [], retry=0 2024-12-02T15:44:21,579 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41981,DS-e3a544b7-2773-454d-9f73-a3c611c95004,DISK] 2024-12-02T15:44:21,579 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42823,DS-2c27ba71-8b49-46d3-8c80-bdf40b0f3bda,DISK] 2024-12-02T15:44:21,579 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40719,DS-9fd6a521-3c0d-4bbf-81e7-a76c42cca598,DISK] 2024-12-02T15:44:21,587 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/WALs/e1c0b6c2739c,34903,1733154259517/e1c0b6c2739c%2C34903%2C1733154259517.meta.1733154261554.meta 2024-12-02T15:44:21,588 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34621:34621),(127.0.0.1/127.0.0.1:41383:41383),(127.0.0.1/127.0.0.1:42411:42411)] 2024-12-02T15:44:21,588 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T15:44:21,590 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T15:44:21,593 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T15:44:21,598 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T15:44:21,601 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T15:44:21,602 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:21,602 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T15:44:21,602 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T15:44:21,605 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T15:44:21,607 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T15:44:21,607 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:21,608 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:21,608 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T15:44:21,610 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T15:44:21,610 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:21,611 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:21,611 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T15:44:21,612 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T15:44:21,612 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:21,613 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:21,613 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T15:44:21,615 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T15:44:21,615 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:21,616 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:21,616 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T15:44:21,618 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740 2024-12-02T15:44:21,621 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740 2024-12-02T15:44:21,624 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T15:44:21,624 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T15:44:21,625 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T15:44:21,627 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T15:44:21,629 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60562698, jitterRate=-0.09754547476768494}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T15:44:21,629 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T15:44:21,630 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733154261603Writing region info on filesystem at 1733154261603Initializing all the Stores at 1733154261605 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154261605Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154261605Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154261605Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154261605Cleaning up temporary data from old regions at 1733154261624 (+19 ms)Running coprocessor post-open hooks at 1733154261629 (+5 ms)Region opened successfully at 1733154261630 (+1 ms) 2024-12-02T15:44:21,637 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733154261525 2024-12-02T15:44:21,648 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T15:44:21,648 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T15:44:21,650 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:21,652 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as e1c0b6c2739c,34903,1733154259517, state=OPEN 2024-12-02T15:44:21,660 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T15:44:21,660 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T15:44:21,660 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T15:44:21,661 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:21,661 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:21,660 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T15:44:21,661 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:21,661 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:21,661 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:21,666 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T15:44:21,666 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=e1c0b6c2739c,34903,1733154259517 in 304 msec 2024-12-02T15:44:21,673 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T15:44:21,674 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 795 msec 2024-12-02T15:44:21,675 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T15:44:21,676 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T15:44:21,693 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T15:44:21,694 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=e1c0b6c2739c,34903,1733154259517, seqNum=-1] 2024-12-02T15:44:21,713 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T15:44:21,715 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59757, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T15:44:21,742 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1210 sec 2024-12-02T15:44:21,742 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733154261742, completionTime=-1 2024-12-02T15:44:21,744 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-02T15:44:21,745 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T15:44:21,769 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-02T15:44:21,769 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733154321769 2024-12-02T15:44:21,769 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733154381769 2024-12-02T15:44:21,769 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 24 msec 2024-12-02T15:44:21,771 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,43127,1733154258626-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,772 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,43127,1733154258626-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,772 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,43127,1733154258626-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,773 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-e1c0b6c2739c:43127, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,774 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,774 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,781 DEBUG [master/e1c0b6c2739c:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T15:44:21,803 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.133sec 2024-12-02T15:44:21,805 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T15:44:21,806 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T15:44:21,807 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T15:44:21,807 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T15:44:21,807 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T15:44:21,808 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,43127,1733154258626-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T15:44:21,808 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,43127,1733154258626-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T15:44:21,817 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T15:44:21,817 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T15:44:21,818 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,43127,1733154258626-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:21,829 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3ad5c43f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T15:44:21,835 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-02T15:44:21,835 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-02T15:44:21,840 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request e1c0b6c2739c,43127,-1 for getting cluster id 2024-12-02T15:44:21,844 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T15:44:21,852 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'c866859f-eef4-41fb-926d-a3811ca09816' 2024-12-02T15:44:21,856 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T15:44:21,856 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "c866859f-eef4-41fb-926d-a3811ca09816" 2024-12-02T15:44:21,857 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@48edb6df, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T15:44:21,857 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [e1c0b6c2739c,43127,-1] 2024-12-02T15:44:21,860 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T15:44:21,863 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:21,864 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:57562, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T15:44:21,868 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6b1ff74e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T15:44:21,868 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T15:44:21,877 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=e1c0b6c2739c,34903,1733154259517, seqNum=-1] 2024-12-02T15:44:21,877 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T15:44:21,880 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46502, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T15:44:21,917 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:21,923 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-02T15:44:21,929 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:21,932 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@664adf48 2024-12-02T15:44:21,933 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T15:44:21,936 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:57578, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T15:44:21,944 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T15:44:21,960 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-02T15:44:21,963 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T15:44:21,966 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-02T15:44:21,966 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:21,969 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T15:44:21,976 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T15:44:22,000 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:22,000 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:22,009 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:36594 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:41981:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36594 dst: /127.0.0.1:41981 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:22,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-02T15:44:22,019 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:22,021 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => c544a3c3d134c05c684dfb83e0b89847, NAME => 'TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f 2024-12-02T15:44:22,028 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:22,028 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:22,038 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:36612 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:41981:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36612 dst: /127.0.0.1:41981 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:22,043 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-02T15:44:22,044 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:22,045 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:22,045 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing c544a3c3d134c05c684dfb83e0b89847, disabling compactions & flushes 2024-12-02T15:44:22,045 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:22,045 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:22,045 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. after waiting 0 ms 2024-12-02T15:44:22,045 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:22,045 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:22,045 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for c544a3c3d134c05c684dfb83e0b89847: Waiting for close lock at 1733154262045Disabling compacts and flushes for region at 1733154262045Disabling writes for close at 1733154262045Writing region close event to WAL at 1733154262045Closed at 1733154262045 2024-12-02T15:44:22,048 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T15:44:22,053 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733154262048"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733154262048"}]},"ts":"1733154262048"} 2024-12-02T15:44:22,058 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-02T15:44:22,061 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T15:44:22,064 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733154262061"}]},"ts":"1733154262061"} 2024-12-02T15:44:22,069 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-02T15:44:22,071 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {e1c0b6c2739c=0} racks are {/default-rack=0} 2024-12-02T15:44:22,074 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-02T15:44:22,075 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-02T15:44:22,075 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-02T15:44:22,075 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-02T15:44:22,075 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-02T15:44:22,075 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-02T15:44:22,075 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-02T15:44:22,075 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-02T15:44:22,075 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-02T15:44:22,075 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-02T15:44:22,077 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=c544a3c3d134c05c684dfb83e0b89847, ASSIGN}] 2024-12-02T15:44:22,079 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=c544a3c3d134c05c684dfb83e0b89847, ASSIGN 2024-12-02T15:44:22,081 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=c544a3c3d134c05c684dfb83e0b89847, ASSIGN; state=OFFLINE, location=e1c0b6c2739c,38775,1733154259568; forceNewPlan=false, retain=false 2024-12-02T15:44:22,094 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T15:44:22,234 INFO [e1c0b6c2739c:43127 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-02T15:44:22,234 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=c544a3c3d134c05c684dfb83e0b89847, regionState=OPENING, regionLocation=e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:22,238 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=c544a3c3d134c05c684dfb83e0b89847, ASSIGN because future has completed 2024-12-02T15:44:22,239 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure c544a3c3d134c05c684dfb83e0b89847, server=e1c0b6c2739c,38775,1733154259568}] 2024-12-02T15:44:22,304 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T15:44:22,394 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T15:44:22,396 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49197, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T15:44:22,405 INFO [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:22,406 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => c544a3c3d134c05c684dfb83e0b89847, NAME => 'TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847.', STARTKEY => '', ENDKEY => ''} 2024-12-02T15:44:22,406 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,406 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:22,407 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,407 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,411 INFO [StoreOpener-c544a3c3d134c05c684dfb83e0b89847-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,413 INFO [StoreOpener-c544a3c3d134c05c684dfb83e0b89847-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region c544a3c3d134c05c684dfb83e0b89847 columnFamilyName cf 2024-12-02T15:44:22,413 DEBUG [StoreOpener-c544a3c3d134c05c684dfb83e0b89847-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:22,414 INFO [StoreOpener-c544a3c3d134c05c684dfb83e0b89847-1 {}] regionserver.HStore(327): Store=c544a3c3d134c05c684dfb83e0b89847/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T15:44:22,415 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,416 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,417 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,417 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,418 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,421 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,436 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T15:44:22,437 INFO [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened c544a3c3d134c05c684dfb83e0b89847; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69267981, jitterRate=0.03217335045337677}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T15:44:22,437 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:22,439 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for c544a3c3d134c05c684dfb83e0b89847: Running coprocessor pre-open hook at 1733154262407Writing region info on filesystem at 1733154262407Initializing all the Stores at 1733154262409 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154262409Cleaning up temporary data from old regions at 1733154262418 (+9 ms)Running coprocessor post-open hooks at 1733154262438 (+20 ms)Region opened successfully at 1733154262439 (+1 ms) 2024-12-02T15:44:22,443 INFO [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847., pid=6, masterSystemTime=1733154262393 2024-12-02T15:44:22,451 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:22,451 INFO [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:22,452 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=c544a3c3d134c05c684dfb83e0b89847, regionState=OPEN, openSeqNum=2, regionLocation=e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:22,460 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure c544a3c3d134c05c684dfb83e0b89847, server=e1c0b6c2739c,38775,1733154259568 because future has completed 2024-12-02T15:44:22,462 WARN [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43127 {}] assignment.AssignmentManager(1543): Unable to acquire lock for regionNode state=OPEN, location=e1c0b6c2739c,38775,1733154259568, table=TestHBaseWalOnEC, region=c544a3c3d134c05c684dfb83e0b89847. It is likely that another thread is currently holding the lock. To avoid deadlock, skip execution for now. 2024-12-02T15:44:22,470 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T15:44:22,471 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure c544a3c3d134c05c684dfb83e0b89847, server=e1c0b6c2739c,38775,1733154259568 in 224 msec 2024-12-02T15:44:22,477 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T15:44:22,477 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=c544a3c3d134c05c684dfb83e0b89847, ASSIGN in 394 msec 2024-12-02T15:44:22,479 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T15:44:22,480 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733154262479"}]},"ts":"1733154262479"} 2024-12-02T15:44:22,484 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-02T15:44:22,495 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T15:44:22,500 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 547 msec 2024-12-02T15:44:22,614 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T15:44:22,615 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-02T15:44:22,615 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-02T15:44:22,616 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T15:44:22,623 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-02T15:44:22,623 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T15:44:22,624 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-02T15:44:22,636 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847., hostname=e1c0b6c2739c,38775,1733154259568, seqNum=2] 2024-12-02T15:44:22,638 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T15:44:22,640 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41984, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T15:44:22,650 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestHBaseWalOnEC 2024-12-02T15:44:22,661 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-02T15:44:22,664 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T15:44:22,664 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-02T15:44:22,667 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T15:44:22,669 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T15:44:22,775 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T15:44:22,847 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=38775 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-02T15:44:22,848 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:22,853 INFO [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing c544a3c3d134c05c684dfb83e0b89847 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-02T15:44:22,976 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847/.tmp/cf/492c75a9284846978e6b8fb64f401722 is 36, key is row/cf:cq/1733154262641/Put/seqid=0 2024-12-02T15:44:22,987 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T15:44:22,989 WARN [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:22,989 WARN [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:23,008 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1577055753_22 at /127.0.0.1:36640 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:41981:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36640 dst: /127.0.0.1:41981 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:23,020 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-02T15:44:23,023 WARN [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:23,023 INFO [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847/.tmp/cf/492c75a9284846978e6b8fb64f401722 2024-12-02T15:44:23,086 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847/.tmp/cf/492c75a9284846978e6b8fb64f401722 as hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847/cf/492c75a9284846978e6b8fb64f401722 2024-12-02T15:44:23,100 INFO [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847/cf/492c75a9284846978e6b8fb64f401722, entries=1, sequenceid=5, filesize=4.7 K 2024-12-02T15:44:23,115 INFO [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for c544a3c3d134c05c684dfb83e0b89847 in 258ms, sequenceid=5, compaction requested=false 2024-12-02T15:44:23,116 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-02T15:44:23,119 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for c544a3c3d134c05c684dfb83e0b89847: 2024-12-02T15:44:23,119 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:23,121 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-02T15:44:23,125 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-02T15:44:23,146 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-02T15:44:23,146 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 463 msec 2024-12-02T15:44:23,158 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 499 msec 2024-12-02T15:44:23,295 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43127 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T15:44:23,296 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-02T15:44:23,316 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T15:44:23,317 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T15:44:23,317 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:23,324 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:23,326 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:23,326 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T15:44:23,327 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T15:44:23,327 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1111665507, stopped=false 2024-12-02T15:44:23,327 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=e1c0b6c2739c,43127,1733154258626 2024-12-02T15:44:23,369 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:23,369 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:23,369 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:23,369 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:23,369 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:23,369 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:23,369 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:23,370 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:23,370 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:23,370 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:23,371 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:23,372 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:23,373 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T15:44:23,373 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T15:44:23,373 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:23,373 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:23,375 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'e1c0b6c2739c,45141,1733154259408' ***** 2024-12-02T15:44:23,375 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T15:44:23,375 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'e1c0b6c2739c,34903,1733154259517' ***** 2024-12-02T15:44:23,375 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T15:44:23,375 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'e1c0b6c2739c,38775,1733154259568' ***** 2024-12-02T15:44:23,375 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T15:44:23,375 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T15:44:23,375 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T15:44:23,375 INFO [RS:2;e1c0b6c2739c:38775 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T15:44:23,375 INFO [RS:0;e1c0b6c2739c:45141 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T15:44:23,375 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T15:44:23,376 INFO [RS:2;e1c0b6c2739c:38775 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T15:44:23,376 INFO [RS:0;e1c0b6c2739c:45141 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T15:44:23,376 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T15:44:23,376 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(959): stopping server e1c0b6c2739c,45141,1733154259408 2024-12-02T15:44:23,376 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T15:44:23,376 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(3091): Received CLOSE for c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:23,376 INFO [RS:0;e1c0b6c2739c:45141 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;e1c0b6c2739c:45141. 2024-12-02T15:44:23,376 DEBUG [RS:0;e1c0b6c2739c:45141 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:23,376 DEBUG [RS:0;e1c0b6c2739c:45141 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:23,377 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T15:44:23,377 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T15:44:23,377 INFO [RS:1;e1c0b6c2739c:34903 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T15:44:23,377 INFO [RS:1;e1c0b6c2739c:34903 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T15:44:23,377 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(959): stopping server e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:23,377 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(976): stopping server e1c0b6c2739c,45141,1733154259408; all regions closed. 2024-12-02T15:44:23,378 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T15:44:23,378 INFO [RS:1;e1c0b6c2739c:34903 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;e1c0b6c2739c:34903. 2024-12-02T15:44:23,378 DEBUG [RS:1;e1c0b6c2739c:34903 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:23,378 DEBUG [RS:1;e1c0b6c2739c:34903 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:23,378 INFO [regionserver/e1c0b6c2739c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:23,379 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(959): stopping server e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:23,379 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T15:44:23,379 INFO [RS:2;e1c0b6c2739c:38775 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;e1c0b6c2739c:38775. 2024-12-02T15:44:23,379 DEBUG [RS:2;e1c0b6c2739c:38775 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:23,379 DEBUG [RS:2;e1c0b6c2739c:38775 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:23,379 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-02T15:44:23,379 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1325): Online Regions={c544a3c3d134c05c684dfb83e0b89847=TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847.} 2024-12-02T15:44:23,380 DEBUG [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1351): Waiting on c544a3c3d134c05c684dfb83e0b89847 2024-12-02T15:44:23,380 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T15:44:23,380 INFO [regionserver/e1c0b6c2739c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:23,380 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T15:44:23,380 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T15:44:23,381 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T15:44:23,382 INFO [regionserver/e1c0b6c2739c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:23,383 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing c544a3c3d134c05c684dfb83e0b89847, disabling compactions & flushes 2024-12-02T15:44:23,383 INFO [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:23,383 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:23,383 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. after waiting 0 ms 2024-12-02T15:44:23,383 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:23,387 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-02T15:44:23,387 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-02T15:44:23,387 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-02T15:44:23,387 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T15:44:23,389 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T15:44:23,390 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T15:44:23,390 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T15:44:23,390 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T15:44:23,390 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-02T15:44:23,425 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_1073741827_1017 (size=93) 2024-12-02T15:44:23,425 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_1073741827_1017 (size=93) 2024-12-02T15:44:23,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_1073741827_1017 (size=93) 2024-12-02T15:44:23,441 DEBUG [RS:0;e1c0b6c2739c:45141 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/oldWALs 2024-12-02T15:44:23,441 INFO [RS:0;e1c0b6c2739c:45141 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL e1c0b6c2739c%2C45141%2C1733154259408:(num 1733154261236) 2024-12-02T15:44:23,442 DEBUG [RS:0;e1c0b6c2739c:45141 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:23,442 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:23,442 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T15:44:23,442 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.ChoreService(370): Chore service for: regionserver/e1c0b6c2739c:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-02T15:44:23,442 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T15:44:23,442 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T15:44:23,442 INFO [regionserver/e1c0b6c2739c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T15:44:23,442 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T15:44:23,442 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T15:44:23,443 INFO [RS:0;e1c0b6c2739c:45141 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:45141 2024-12-02T15:44:23,452 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T15:44:23,452 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/e1c0b6c2739c,45141,1733154259408 2024-12-02T15:44:23,453 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T15:44:23,454 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [e1c0b6c2739c,45141,1733154259408] 2024-12-02T15:44:23,461 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/e1c0b6c2739c,45141,1733154259408 already deleted, retry=false 2024-12-02T15:44:23,461 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; e1c0b6c2739c,45141,1733154259408 expired; onlineServers=2 2024-12-02T15:44:23,499 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/info/cc8d0e7a75084926a5ccc3b6c8e9fb65 is 153, key is TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847./info:regioninfo/1733154262451/Put/seqid=0 2024-12-02T15:44:23,508 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:23,508 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:23,513 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/default/TestHBaseWalOnEC/c544a3c3d134c05c684dfb83e0b89847/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-02T15:44:23,519 INFO [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:23,520 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for c544a3c3d134c05c684dfb83e0b89847: Waiting for close lock at 1733154263383Running coprocessor pre-close hooks at 1733154263383Disabling compacts and flushes for region at 1733154263383Disabling writes for close at 1733154263383Writing region close event to WAL at 1733154263431 (+48 ms)Running coprocessor post-close hooks at 1733154263515 (+84 ms)Closed at 1733154263519 (+4 ms) 2024-12-02T15:44:23,520 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733154261938.c544a3c3d134c05c684dfb83e0b89847. 2024-12-02T15:44:23,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775741_1008 (size=1189) 2024-12-02T15:44:23,543 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_298469757_22 at /127.0.0.1:35610 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:42823:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35610 dst: /127.0.0.1:42823 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:23,547 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775772_1004 (size=42) 2024-12-02T15:44:23,549 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775757_1006 (size=196) 2024-12-02T15:44:23,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_-9223372036854775740_1008 (size=1189) 2024-12-02T15:44:23,558 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775756_1006 (size=196) 2024-12-02T15:44:23,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775725_1010 (size=34) 2024-12-02T15:44:23,564 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_-9223372036854775724_1010 (size=34) 2024-12-02T15:44:23,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-02T15:44:23,568 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:23,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775773_1004 (size=42) 2024-12-02T15:44:23,568 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/info/cc8d0e7a75084926a5ccc3b6c8e9fb65 2024-12-02T15:44:23,570 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:23,570 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45141-0x1019800de120001, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:23,571 INFO [RS:0;e1c0b6c2739c:45141 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T15:44:23,571 INFO [RS:0;e1c0b6c2739c:45141 {}] regionserver.HRegionServer(1031): Exiting; stopping=e1c0b6c2739c,45141,1733154259408; zookeeper connection closed. 2024-12-02T15:44:23,573 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@3415799c {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@3415799c 2024-12-02T15:44:23,580 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(976): stopping server e1c0b6c2739c,38775,1733154259568; all regions closed. 2024-12-02T15:44:23,587 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-02T15:44:23,604 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_1073741828_1018 (size=1298) 2024-12-02T15:44:23,604 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_1073741828_1018 (size=1298) 2024-12-02T15:44:23,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_1073741828_1018 (size=1298) 2024-12-02T15:44:23,613 DEBUG [RS:2;e1c0b6c2739c:38775 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/oldWALs 2024-12-02T15:44:23,613 INFO [RS:2;e1c0b6c2739c:38775 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL e1c0b6c2739c%2C38775%2C1733154259568:(num 1733154261242) 2024-12-02T15:44:23,613 DEBUG [RS:2;e1c0b6c2739c:38775 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:23,613 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:23,617 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T15:44:23,617 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.ChoreService(370): Chore service for: regionserver/e1c0b6c2739c:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-02T15:44:23,618 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T15:44:23,618 INFO [regionserver/e1c0b6c2739c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T15:44:23,618 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T15:44:23,618 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T15:44:23,618 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T15:44:23,618 INFO [RS:2;e1c0b6c2739c:38775 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:38775 2024-12-02T15:44:23,627 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/e1c0b6c2739c,38775,1733154259568 2024-12-02T15:44:23,627 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T15:44:23,627 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T15:44:23,635 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/ns/a3880cc28ff6441e86f6e3f624421f3d is 43, key is default/ns:d/1733154261719/Put/seqid=0 2024-12-02T15:44:23,636 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [e1c0b6c2739c,38775,1733154259568] 2024-12-02T15:44:23,639 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:23,639 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:23,644 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/e1c0b6c2739c,38775,1733154259568 already deleted, retry=false 2024-12-02T15:44:23,644 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; e1c0b6c2739c,38775,1733154259568 expired; onlineServers=1 2024-12-02T15:44:23,647 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_298469757_22 at /127.0.0.1:35666 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:42823:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35666 dst: /127.0.0.1:42823 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:23,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-02T15:44:23,660 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:23,660 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/ns/a3880cc28ff6441e86f6e3f624421f3d 2024-12-02T15:44:23,694 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/table/346493367a5e486d87ae66094f36d15a is 52, key is TestHBaseWalOnEC/table:state/1733154262479/Put/seqid=0 2024-12-02T15:44:23,697 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:23,697 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:23,720 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_298469757_22 at /127.0.0.1:35694 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:42823:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35694 dst: /127.0.0.1:42823 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:23,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-02T15:44:23,736 INFO [RS:2;e1c0b6c2739c:38775 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T15:44:23,736 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:23,736 INFO [RS:2;e1c0b6c2739c:38775 {}] regionserver.HRegionServer(1031): Exiting; stopping=e1c0b6c2739c,38775,1733154259568; zookeeper connection closed. 2024-12-02T15:44:23,736 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38775-0x1019800de120003, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:23,739 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@3a3a568c {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@3a3a568c 2024-12-02T15:44:23,788 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-02T15:44:23,988 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-02T15:44:24,072 INFO [regionserver/e1c0b6c2739c:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-02T15:44:24,072 INFO [regionserver/e1c0b6c2739c:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-02T15:44:24,135 WARN [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:24,135 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/table/346493367a5e486d87ae66094f36d15a 2024-12-02T15:44:24,151 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/info/cc8d0e7a75084926a5ccc3b6c8e9fb65 as hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/info/cc8d0e7a75084926a5ccc3b6c8e9fb65 2024-12-02T15:44:24,164 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/info/cc8d0e7a75084926a5ccc3b6c8e9fb65, entries=10, sequenceid=11, filesize=6.5 K 2024-12-02T15:44:24,166 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/ns/a3880cc28ff6441e86f6e3f624421f3d as hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/ns/a3880cc28ff6441e86f6e3f624421f3d 2024-12-02T15:44:24,179 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/ns/a3880cc28ff6441e86f6e3f624421f3d, entries=2, sequenceid=11, filesize=5.0 K 2024-12-02T15:44:24,182 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/.tmp/table/346493367a5e486d87ae66094f36d15a as hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/table/346493367a5e486d87ae66094f36d15a 2024-12-02T15:44:24,188 DEBUG [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-02T15:44:24,200 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/table/346493367a5e486d87ae66094f36d15a, entries=2, sequenceid=11, filesize=5.1 K 2024-12-02T15:44:24,202 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 812ms, sequenceid=11, compaction requested=false 2024-12-02T15:44:24,203 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T15:44:24,228 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-02T15:44:24,230 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T15:44:24,230 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T15:44:24,230 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733154263387Running coprocessor pre-close hooks at 1733154263387Disabling compacts and flushes for region at 1733154263387Disabling writes for close at 1733154263390 (+3 ms)Obtaining lock to block concurrent updates at 1733154263390Preparing flush snapshotting stores in 1588230740 at 1733154263391 (+1 ms)Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733154263393 (+2 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733154263399 (+6 ms)Flushing 1588230740/info: creating writer at 1733154263400 (+1 ms)Flushing 1588230740/info: appending metadata at 1733154263494 (+94 ms)Flushing 1588230740/info: closing flushed file at 1733154263494Flushing 1588230740/ns: creating writer at 1733154263592 (+98 ms)Flushing 1588230740/ns: appending metadata at 1733154263631 (+39 ms)Flushing 1588230740/ns: closing flushed file at 1733154263631Flushing 1588230740/table: creating writer at 1733154263671 (+40 ms)Flushing 1588230740/table: appending metadata at 1733154263693 (+22 ms)Flushing 1588230740/table: closing flushed file at 1733154263693Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3bdb9d9d: reopening flushed file at 1733154264149 (+456 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@269cf99c: reopening flushed file at 1733154264164 (+15 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3b9f7027: reopening flushed file at 1733154264180 (+16 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 812ms, sequenceid=11, compaction requested=false at 1733154264203 (+23 ms)Writing region close event to WAL at 1733154264206 (+3 ms)Running coprocessor post-close hooks at 1733154264229 (+23 ms)Closed at 1733154264230 (+1 ms) 2024-12-02T15:44:24,231 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T15:44:24,389 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(976): stopping server e1c0b6c2739c,34903,1733154259517; all regions closed. 2024-12-02T15:44:24,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_1073741829_1019 (size=2751) 2024-12-02T15:44:24,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_1073741829_1019 (size=2751) 2024-12-02T15:44:24,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_1073741829_1019 (size=2751) 2024-12-02T15:44:24,396 DEBUG [RS:1;e1c0b6c2739c:34903 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/oldWALs 2024-12-02T15:44:24,396 INFO [RS:1;e1c0b6c2739c:34903 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL e1c0b6c2739c%2C34903%2C1733154259517.meta:.meta(num 1733154261554) 2024-12-02T15:44:24,406 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_1073741826_1016 (size=93) 2024-12-02T15:44:24,406 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_1073741826_1016 (size=93) 2024-12-02T15:44:24,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_1073741826_1016 (size=93) 2024-12-02T15:44:24,414 DEBUG [RS:1;e1c0b6c2739c:34903 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/oldWALs 2024-12-02T15:44:24,414 INFO [RS:1;e1c0b6c2739c:34903 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL e1c0b6c2739c%2C34903%2C1733154259517:(num 1733154261229) 2024-12-02T15:44:24,414 DEBUG [RS:1;e1c0b6c2739c:34903 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:24,414 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:24,414 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T15:44:24,414 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.ChoreService(370): Chore service for: regionserver/e1c0b6c2739c:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-02T15:44:24,414 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T15:44:24,414 INFO [regionserver/e1c0b6c2739c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T15:44:24,415 INFO [RS:1;e1c0b6c2739c:34903 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:34903 2024-12-02T15:44:24,427 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T15:44:24,427 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T15:44:24,427 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/e1c0b6c2739c,34903,1733154259517 2024-12-02T15:44:24,427 ERROR [Time-limited test-EventThread {}] zookeeper.ClientCnxn$EventThread(581): Error while calling watcher. java.util.concurrent.RejectedExecutionException: Task org.apache.hadoop.hbase.trace.TraceUtil$$Lambda$377/0x00007f95248f60d8@5bd2c463 rejected from java.util.concurrent.ThreadPoolExecutor@1ed0517d[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 14] at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2065) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:833) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1365) ~[?:?] at java.util.concurrent.Executors$DelegatedExecutorService.execute(Executors.java:721) ~[?:?] at org.apache.hadoop.hbase.zookeeper.ZKWatcher.process(ZKWatcher.java:613) ~[hbase-zookeeper-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:579) ~[zookeeper-3.8.4.jar:3.8.4] at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:554) ~[zookeeper-3.8.4.jar:3.8.4] 2024-12-02T15:44:24,436 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [e1c0b6c2739c,34903,1733154259517] 2024-12-02T15:44:24,443 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/e1c0b6c2739c,34903,1733154259517 already deleted, retry=false 2024-12-02T15:44:24,444 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; e1c0b6c2739c,34903,1733154259517 expired; onlineServers=0 2024-12-02T15:44:24,444 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'e1c0b6c2739c,43127,1733154258626' ***** 2024-12-02T15:44:24,444 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T15:44:24,444 INFO [M:0;e1c0b6c2739c:43127 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T15:44:24,444 INFO [M:0;e1c0b6c2739c:43127 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T15:44:24,444 DEBUG [M:0;e1c0b6c2739c:43127 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T15:44:24,444 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T15:44:24,444 DEBUG [M:0;e1c0b6c2739c:43127 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T15:44:24,445 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.large.0-1733154260743 {}] cleaner.HFileCleaner(306): Exit Thread[master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.large.0-1733154260743,5,FailOnTimeoutGroup] 2024-12-02T15:44:24,444 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.small.0-1733154260747 {}] cleaner.HFileCleaner(306): Exit Thread[master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.small.0-1733154260747,5,FailOnTimeoutGroup] 2024-12-02T15:44:24,445 INFO [M:0;e1c0b6c2739c:43127 {}] hbase.ChoreService(370): Chore service for: master/e1c0b6c2739c:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T15:44:24,445 INFO [M:0;e1c0b6c2739c:43127 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T15:44:24,445 DEBUG [M:0;e1c0b6c2739c:43127 {}] master.HMaster(1795): Stopping service threads 2024-12-02T15:44:24,445 INFO [M:0;e1c0b6c2739c:43127 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T15:44:24,445 INFO [M:0;e1c0b6c2739c:43127 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T15:44:24,446 INFO [M:0;e1c0b6c2739c:43127 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T15:44:24,446 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T15:44:24,452 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T15:44:24,452 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:24,452 DEBUG [M:0;e1c0b6c2739c:43127 {}] zookeeper.ZKUtil(347): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T15:44:24,452 WARN [M:0;e1c0b6c2739c:43127 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T15:44:24,453 INFO [M:0;e1c0b6c2739c:43127 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/.lastflushedseqids 2024-12-02T15:44:24,467 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:24,467 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:24,471 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:34324 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:40719:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34324 dst: /127.0.0.1:40719 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:24,488 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-02T15:44:24,490 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:24,490 INFO [M:0;e1c0b6c2739c:43127 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T15:44:24,491 INFO [M:0;e1c0b6c2739c:43127 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T15:44:24,491 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T15:44:24,491 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:24,491 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:24,491 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T15:44:24,491 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:24,491 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.84 KB heapSize=34.13 KB 2024-12-02T15:44:24,513 DEBUG [M:0;e1c0b6c2739c:43127 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/715bb9a0ff9945e1a5d504870ee866f9 is 82, key is hbase:meta,,1/info:regioninfo/1733154261649/Put/seqid=0 2024-12-02T15:44:24,515 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:24,515 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:24,518 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:36684 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:41981:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36684 dst: /127.0.0.1:41981 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:24,524 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-02T15:44:24,524 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:24,524 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/715bb9a0ff9945e1a5d504870ee866f9 2024-12-02T15:44:24,536 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:24,536 INFO [RS:1;e1c0b6c2739c:34903 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T15:44:24,536 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34903-0x1019800de120002, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:24,536 INFO [RS:1;e1c0b6c2739c:34903 {}] regionserver.HRegionServer(1031): Exiting; stopping=e1c0b6c2739c,34903,1733154259517; zookeeper connection closed. 2024-12-02T15:44:24,536 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5267f5db {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5267f5db 2024-12-02T15:44:24,536 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-02T15:44:24,548 DEBUG [M:0;e1c0b6c2739c:43127 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/935c8ee6525a4d59bcd4dae98813d0b7 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733154262498/Put/seqid=0 2024-12-02T15:44:24,550 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:24,551 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:24,558 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:35716 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:42823:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35716 dst: /127.0.0.1:42823 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:24,562 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775552_1037 (size=6440) 2024-12-02T15:44:24,563 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:24,563 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.15 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/935c8ee6525a4d59bcd4dae98813d0b7 2024-12-02T15:44:24,597 DEBUG [M:0;e1c0b6c2739c:43127 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/b078b22b4f474c4fa7a76744e1544bee is 69, key is e1c0b6c2739c,34903,1733154259517/rs:state/1733154260880/Put/seqid=0 2024-12-02T15:44:24,600 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:24,600 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-02T15:44:24,603 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1158563463_22 at /127.0.0.1:35746 [Receiving block BP-948202177-172.17.0.3-1733154254750:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:42823:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35746 dst: /127.0.0.1:42823 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T15:44:24,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-02T15:44:24,607 WARN [M:0;e1c0b6c2739c:43127 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-02T15:44:24,607 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/b078b22b4f474c4fa7a76744e1544bee 2024-12-02T15:44:24,617 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/715bb9a0ff9945e1a5d504870ee866f9 as hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/715bb9a0ff9945e1a5d504870ee866f9 2024-12-02T15:44:24,633 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/715bb9a0ff9945e1a5d504870ee866f9, entries=8, sequenceid=72, filesize=5.5 K 2024-12-02T15:44:24,634 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/935c8ee6525a4d59bcd4dae98813d0b7 as hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/935c8ee6525a4d59bcd4dae98813d0b7 2024-12-02T15:44:24,643 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/935c8ee6525a4d59bcd4dae98813d0b7, entries=8, sequenceid=72, filesize=6.3 K 2024-12-02T15:44:24,645 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/b078b22b4f474c4fa7a76744e1544bee as hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/b078b22b4f474c4fa7a76744e1544bee 2024-12-02T15:44:24,657 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/b078b22b4f474c4fa7a76744e1544bee, entries=3, sequenceid=72, filesize=5.2 K 2024-12-02T15:44:24,659 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 168ms, sequenceid=72, compaction requested=false 2024-12-02T15:44:24,663 INFO [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:24,663 DEBUG [M:0;e1c0b6c2739c:43127 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733154264491Disabling compacts and flushes for region at 1733154264491Disabling writes for close at 1733154264491Obtaining lock to block concurrent updates at 1733154264491Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733154264491Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27480, getHeapSize=34880, getOffHeapSize=0, getCellsCount=85 at 1733154264492 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733154264493 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733154264494 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733154264513 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733154264513Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733154264533 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733154264548 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733154264548Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733154264572 (+24 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733154264597 (+25 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733154264597Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@74f3d03f: reopening flushed file at 1733154264615 (+18 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@63b8c045: reopening flushed file at 1733154264633 (+18 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7d35b1bc: reopening flushed file at 1733154264644 (+11 ms)Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 168ms, sequenceid=72, compaction requested=false at 1733154264659 (+15 ms)Writing region close event to WAL at 1733154264663 (+4 ms)Closed at 1733154264663 2024-12-02T15:44:24,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42823 is added to blk_1073741825_1011 (size=32683) 2024-12-02T15:44:24,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40719 is added to blk_1073741825_1011 (size=32683) 2024-12-02T15:44:24,672 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41981 is added to blk_1073741825_1011 (size=32683) 2024-12-02T15:44:24,673 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T15:44:24,673 INFO [M:0;e1c0b6c2739c:43127 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T15:44:24,673 INFO [M:0;e1c0b6c2739c:43127 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:43127 2024-12-02T15:44:24,673 INFO [M:0;e1c0b6c2739c:43127 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T15:44:24,786 INFO [M:0;e1c0b6c2739c:43127 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T15:44:24,786 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:24,786 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43127-0x1019800de120000, quorum=127.0.0.1:55083, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:24,823 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7f750918{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:24,825 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@86bf2a7{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T15:44:24,825 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T15:44:24,825 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1023f385{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T15:44:24,825 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7e1f796{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir/,STOPPED} 2024-12-02T15:44:24,831 WARN [BP-948202177-172.17.0.3-1733154254750 heartbeating to localhost/127.0.0.1:45117 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T15:44:24,831 WARN [BP-948202177-172.17.0.3-1733154254750 heartbeating to localhost/127.0.0.1:45117 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-948202177-172.17.0.3-1733154254750 (Datanode Uuid 98925a14-9d17-469e-9bda-96a2c4e78ae9) service to localhost/127.0.0.1:45117 2024-12-02T15:44:24,833 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data5/current/BP-948202177-172.17.0.3-1733154254750 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:24,833 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data6/current/BP-948202177-172.17.0.3-1733154254750 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:24,833 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T15:44:24,833 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T15:44:24,834 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T15:44:24,841 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@26b068f7{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:24,842 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5739b847{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T15:44:24,842 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T15:44:24,842 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2c2c5be{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T15:44:24,842 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4a91ec1e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir/,STOPPED} 2024-12-02T15:44:24,844 WARN [BP-948202177-172.17.0.3-1733154254750 heartbeating to localhost/127.0.0.1:45117 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T15:44:24,844 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T15:44:24,844 WARN [BP-948202177-172.17.0.3-1733154254750 heartbeating to localhost/127.0.0.1:45117 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-948202177-172.17.0.3-1733154254750 (Datanode Uuid 972a9d42-1c27-4269-a769-8cd739344f28) service to localhost/127.0.0.1:45117 2024-12-02T15:44:24,844 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T15:44:24,846 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data3/current/BP-948202177-172.17.0.3-1733154254750 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:24,846 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data4/current/BP-948202177-172.17.0.3-1733154254750 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:24,847 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T15:44:24,856 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7e705dc8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:24,856 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@2ad1569e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T15:44:24,856 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T15:44:24,856 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@17f1c7fc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T15:44:24,856 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@32fec40a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir/,STOPPED} 2024-12-02T15:44:24,858 WARN [BP-948202177-172.17.0.3-1733154254750 heartbeating to localhost/127.0.0.1:45117 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T15:44:24,858 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T15:44:24,858 WARN [BP-948202177-172.17.0.3-1733154254750 heartbeating to localhost/127.0.0.1:45117 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-948202177-172.17.0.3-1733154254750 (Datanode Uuid 93edeccc-6637-4a92-968d-799c841d9fdb) service to localhost/127.0.0.1:45117 2024-12-02T15:44:24,858 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T15:44:24,859 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data1/current/BP-948202177-172.17.0.3-1733154254750 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:24,859 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/cluster_a2bde83d-733a-61a3-fb27-49a852e1b05d/data/data2/current/BP-948202177-172.17.0.3-1733154254750 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:24,859 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T15:44:24,872 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@76e4c45c{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T15:44:24,872 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T15:44:24,873 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T15:44:24,873 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T15:44:24,873 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir/,STOPPED} 2024-12-02T15:44:24,881 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T15:44:24,910 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T15:44:24,923 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=91 (was 161), OpenFileDescriptor=439 (was 393) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=636 (was 652), ProcessCount=11 (was 11), AvailableMemoryMB=9607 (was 8818) - AvailableMemoryMB LEAK? - 2024-12-02T15:44:24,930 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=91, OpenFileDescriptor=439, MaxFileDescriptor=1048576, SystemLoadAverage=636, ProcessCount=11, AvailableMemoryMB=9607 2024-12-02T15:44:24,930 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T15:44:24,930 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.log.dir so I do NOT create it in target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa 2024-12-02T15:44:24,930 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/6e3fc805-4e51-5a4c-d469-394ef0363313/hadoop.tmp.dir so I do NOT create it in target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa 2024-12-02T15:44:24,930 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd, deleteOnExit=true 2024-12-02T15:44:24,930 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-02T15:44:24,931 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/test.cache.data in system properties and HBase conf 2024-12-02T15:44:24,931 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T15:44:24,931 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir in system properties and HBase conf 2024-12-02T15:44:24,931 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T15:44:24,931 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T15:44:24,931 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T15:44:24,931 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T15:44:24,931 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T15:44:24,931 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/nfs.dump.dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/java.io.tmpdir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T15:44:24,932 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T15:44:25,176 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:25,181 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T15:44:25,182 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T15:44:25,182 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T15:44:25,182 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T15:44:25,183 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:25,184 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@a49b909{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir/,AVAILABLE} 2024-12-02T15:44:25,184 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@56aa9d3b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T15:44:25,283 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7aaeb6cf{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/java.io.tmpdir/jetty-localhost-45511-hadoop-hdfs-3_4_1-tests_jar-_-any-14470399913139071827/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T15:44:25,283 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@686c9dd5{HTTP/1.1, (http/1.1)}{localhost:45511} 2024-12-02T15:44:25,283 INFO [Time-limited test {}] server.Server(415): Started @12946ms 2024-12-02T15:44:25,467 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:25,470 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T15:44:25,471 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T15:44:25,471 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T15:44:25,471 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T15:44:25,472 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2807f8c2{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir/,AVAILABLE} 2024-12-02T15:44:25,472 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@61a92fea{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T15:44:25,603 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@38e5384{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/java.io.tmpdir/jetty-localhost-43477-hadoop-hdfs-3_4_1-tests_jar-_-any-113703674476150663/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:25,604 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@7d6118e0{HTTP/1.1, (http/1.1)}{localhost:43477} 2024-12-02T15:44:25,604 INFO [Time-limited test {}] server.Server(415): Started @13266ms 2024-12-02T15:44:25,606 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T15:44:25,649 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:25,652 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T15:44:25,653 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T15:44:25,653 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T15:44:25,653 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T15:44:25,654 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5b4297c4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir/,AVAILABLE} 2024-12-02T15:44:25,654 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@bb1336{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T15:44:25,764 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6e5e4927{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/java.io.tmpdir/jetty-localhost-42107-hadoop-hdfs-3_4_1-tests_jar-_-any-1268411345164501047/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:25,765 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1768a8c1{HTTP/1.1, (http/1.1)}{localhost:42107} 2024-12-02T15:44:25,765 INFO [Time-limited test {}] server.Server(415): Started @13427ms 2024-12-02T15:44:25,767 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T15:44:25,810 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T15:44:25,814 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T15:44:25,815 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T15:44:25,815 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T15:44:25,815 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T15:44:25,816 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6e0095f0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir/,AVAILABLE} 2024-12-02T15:44:25,816 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@38da8210{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T15:44:25,923 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@bff0a43{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/java.io.tmpdir/jetty-localhost-33759-hadoop-hdfs-3_4_1-tests_jar-_-any-15992474368212728540/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:25,925 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@19dff04d{HTTP/1.1, (http/1.1)}{localhost:33759} 2024-12-02T15:44:25,926 INFO [Time-limited test {}] server.Server(415): Started @13588ms 2024-12-02T15:44:25,928 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T15:44:26,224 WARN [Thread-555 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data1/current/BP-788991653-172.17.0.3-1733154264965/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:26,224 WARN [Thread-556 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data2/current/BP-788991653-172.17.0.3-1733154264965/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:26,269 WARN [Thread-498 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T15:44:26,274 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdcd6ffeab4786808 with lease ID 0x13c92030e8d6831c: Processing first storage report for DS-703e7f1d-69ac-4b4f-b7dd-9072c8d99a65 from datanode DatanodeRegistration(127.0.0.1:41631, datanodeUuid=55b82f5a-737b-47d4-a608-9da47531bbcb, infoPort=43789, infoSecurePort=0, ipcPort=42657, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965) 2024-12-02T15:44:26,274 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdcd6ffeab4786808 with lease ID 0x13c92030e8d6831c: from storage DS-703e7f1d-69ac-4b4f-b7dd-9072c8d99a65 node DatanodeRegistration(127.0.0.1:41631, datanodeUuid=55b82f5a-737b-47d4-a608-9da47531bbcb, infoPort=43789, infoSecurePort=0, ipcPort=42657, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:26,274 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdcd6ffeab4786808 with lease ID 0x13c92030e8d6831c: Processing first storage report for DS-d3b97b24-8185-4b3f-b74f-0848bf1b35fb from datanode DatanodeRegistration(127.0.0.1:41631, datanodeUuid=55b82f5a-737b-47d4-a608-9da47531bbcb, infoPort=43789, infoSecurePort=0, ipcPort=42657, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965) 2024-12-02T15:44:26,274 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdcd6ffeab4786808 with lease ID 0x13c92030e8d6831c: from storage DS-d3b97b24-8185-4b3f-b74f-0848bf1b35fb node DatanodeRegistration(127.0.0.1:41631, datanodeUuid=55b82f5a-737b-47d4-a608-9da47531bbcb, infoPort=43789, infoSecurePort=0, ipcPort=42657, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:26,529 WARN [Thread-570 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data4/current/BP-788991653-172.17.0.3-1733154264965/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:26,529 WARN [Thread-569 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data3/current/BP-788991653-172.17.0.3-1733154264965/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:26,551 WARN [Thread-521 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T15:44:26,553 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5170428c79551b82 with lease ID 0x13c92030e8d6831d: Processing first storage report for DS-a7cd3327-5948-4ebb-802e-4e892d9f7cf3 from datanode DatanodeRegistration(127.0.0.1:39273, datanodeUuid=9d116beb-8d2b-4bee-b153-cb6b098221bd, infoPort=46117, infoSecurePort=0, ipcPort=46043, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965) 2024-12-02T15:44:26,553 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5170428c79551b82 with lease ID 0x13c92030e8d6831d: from storage DS-a7cd3327-5948-4ebb-802e-4e892d9f7cf3 node DatanodeRegistration(127.0.0.1:39273, datanodeUuid=9d116beb-8d2b-4bee-b153-cb6b098221bd, infoPort=46117, infoSecurePort=0, ipcPort=46043, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:26,553 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5170428c79551b82 with lease ID 0x13c92030e8d6831d: Processing first storage report for DS-c0720398-6421-4242-8d31-120036ff126f from datanode DatanodeRegistration(127.0.0.1:39273, datanodeUuid=9d116beb-8d2b-4bee-b153-cb6b098221bd, infoPort=46117, infoSecurePort=0, ipcPort=46043, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965) 2024-12-02T15:44:26,553 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5170428c79551b82 with lease ID 0x13c92030e8d6831d: from storage DS-c0720398-6421-4242-8d31-120036ff126f node DatanodeRegistration(127.0.0.1:39273, datanodeUuid=9d116beb-8d2b-4bee-b153-cb6b098221bd, infoPort=46117, infoSecurePort=0, ipcPort=46043, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:26,706 WARN [Thread-580 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data5/current/BP-788991653-172.17.0.3-1733154264965/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:26,713 WARN [Thread-581 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data6/current/BP-788991653-172.17.0.3-1733154264965/current, will proceed with Du for space computation calculation, 2024-12-02T15:44:26,730 WARN [Thread-545 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T15:44:26,733 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xed202abdda5f1cd8 with lease ID 0x13c92030e8d6831e: Processing first storage report for DS-d47f47cf-80d8-40fc-b2af-7e3fb31e7ace from datanode DatanodeRegistration(127.0.0.1:36501, datanodeUuid=07fcc5c6-5f77-4614-994c-f6ea8a3d5991, infoPort=44569, infoSecurePort=0, ipcPort=40019, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965) 2024-12-02T15:44:26,733 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xed202abdda5f1cd8 with lease ID 0x13c92030e8d6831e: from storage DS-d47f47cf-80d8-40fc-b2af-7e3fb31e7ace node DatanodeRegistration(127.0.0.1:36501, datanodeUuid=07fcc5c6-5f77-4614-994c-f6ea8a3d5991, infoPort=44569, infoSecurePort=0, ipcPort=40019, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:26,733 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xed202abdda5f1cd8 with lease ID 0x13c92030e8d6831e: Processing first storage report for DS-3fae4906-b50d-45c5-ba5e-b6de2c684282 from datanode DatanodeRegistration(127.0.0.1:36501, datanodeUuid=07fcc5c6-5f77-4614-994c-f6ea8a3d5991, infoPort=44569, infoSecurePort=0, ipcPort=40019, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965) 2024-12-02T15:44:26,733 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xed202abdda5f1cd8 with lease ID 0x13c92030e8d6831e: from storage DS-3fae4906-b50d-45c5-ba5e-b6de2c684282 node DatanodeRegistration(127.0.0.1:36501, datanodeUuid=07fcc5c6-5f77-4614-994c-f6ea8a3d5991, infoPort=44569, infoSecurePort=0, ipcPort=40019, storageInfo=lv=-57;cid=testClusterID;nsid=598845607;c=1733154264965), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T15:44:26,812 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa 2024-12-02T15:44:26,814 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/zookeeper_0, clientPort=62243, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T15:44:26,815 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=62243 2024-12-02T15:44:26,816 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:26,818 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:26,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741825_1001 (size=7) 2024-12-02T15:44:26,839 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741825_1001 (size=7) 2024-12-02T15:44:26,840 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741825_1001 (size=7) 2024-12-02T15:44:26,841 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4 with version=8 2024-12-02T15:44:26,841 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:45117/user/jenkins/test-data/6e7a6c7e-eaeb-49aa-62c6-10e08e94166f/hbase-staging 2024-12-02T15:44:26,843 INFO [Time-limited test {}] client.ConnectionUtils(128): master/e1c0b6c2739c:0 server-side Connection retries=45 2024-12-02T15:44:26,843 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:26,843 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:26,843 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T15:44:26,843 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:26,843 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T15:44:26,843 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T15:44:26,843 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T15:44:26,844 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:42329 2024-12-02T15:44:26,846 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:42329 connecting to ZooKeeper ensemble=127.0.0.1:62243 2024-12-02T15:44:26,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:423290x0, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T15:44:26,905 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:42329-0x101980101330000 connected 2024-12-02T15:44:26,970 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:26,971 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:26,973 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:26,973 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4, hbase.cluster.distributed=false 2024-12-02T15:44:26,975 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T15:44:26,975 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=42329 2024-12-02T15:44:26,976 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=42329 2024-12-02T15:44:26,976 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=42329 2024-12-02T15:44:26,976 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=42329 2024-12-02T15:44:26,976 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=42329 2024-12-02T15:44:26,992 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/e1c0b6c2739c:0 server-side Connection retries=45 2024-12-02T15:44:26,992 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:26,992 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:26,992 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T15:44:26,992 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:26,992 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T15:44:26,992 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T15:44:26,993 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T15:44:26,993 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:35335 2024-12-02T15:44:26,995 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:35335 connecting to ZooKeeper ensemble=127.0.0.1:62243 2024-12-02T15:44:26,997 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:27,000 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:27,014 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:353350x0, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:27,014 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T15:44:27,016 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:353350x0, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T15:44:27,017 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T15:44:27,018 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:353350x0, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T15:44:27,019 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35335-0x101980101330001 connected 2024-12-02T15:44:27,019 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T15:44:27,023 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35335 2024-12-02T15:44:27,023 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35335 2024-12-02T15:44:27,026 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35335 2024-12-02T15:44:27,026 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35335 2024-12-02T15:44:27,027 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35335 2024-12-02T15:44:27,040 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/e1c0b6c2739c:0 server-side Connection retries=45 2024-12-02T15:44:27,040 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:27,040 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:27,040 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T15:44:27,040 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:27,040 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T15:44:27,040 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T15:44:27,040 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T15:44:27,041 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:36719 2024-12-02T15:44:27,043 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36719 connecting to ZooKeeper ensemble=127.0.0.1:62243 2024-12-02T15:44:27,044 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:27,046 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:27,060 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:367190x0, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T15:44:27,060 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36719-0x101980101330002 connected 2024-12-02T15:44:27,061 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:27,061 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T15:44:27,062 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T15:44:27,063 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T15:44:27,065 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T15:44:27,070 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36719 2024-12-02T15:44:27,071 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36719 2024-12-02T15:44:27,072 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36719 2024-12-02T15:44:27,072 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36719 2024-12-02T15:44:27,072 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36719 2024-12-02T15:44:27,087 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/e1c0b6c2739c:0 server-side Connection retries=45 2024-12-02T15:44:27,087 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:27,087 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:27,087 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T15:44:27,087 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T15:44:27,087 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T15:44:27,087 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T15:44:27,087 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T15:44:27,091 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:45273 2024-12-02T15:44:27,093 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:45273 connecting to ZooKeeper ensemble=127.0.0.1:62243 2024-12-02T15:44:27,094 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:27,097 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:27,110 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T15:44:27,118 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:452730x0, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T15:44:27,119 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:452730x0, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:27,119 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T15:44:27,124 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45273-0x101980101330003 connected 2024-12-02T15:44:27,126 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T15:44:27,127 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T15:44:27,131 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T15:44:27,138 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45273 2024-12-02T15:44:27,139 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45273 2024-12-02T15:44:27,139 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45273 2024-12-02T15:44:27,148 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45273 2024-12-02T15:44:27,148 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45273 2024-12-02T15:44:27,162 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;e1c0b6c2739c:42329 2024-12-02T15:44:27,162 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:27,167 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T15:44:27,168 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T15:44:27,170 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:27,170 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:27,170 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:27,170 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:27,171 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:27,178 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T15:44:27,178 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T15:44:27,178 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T15:44:27,178 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,178 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,178 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,179 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,179 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T15:44:27,180 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/e1c0b6c2739c,42329,1733154266842 from backup master directory 2024-12-02T15:44:27,221 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:27,221 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:27,221 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:27,221 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T15:44:27,221 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:27,221 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:27,221 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T15:44:27,231 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/hbase.id] with ID: f68683c7-d3bb-4f82-b24f-82d1f72fbf63 2024-12-02T15:44:27,231 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/.tmp/hbase.id 2024-12-02T15:44:27,423 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741826_1002 (size=42) 2024-12-02T15:44:27,423 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741826_1002 (size=42) 2024-12-02T15:44:27,424 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741826_1002 (size=42) 2024-12-02T15:44:27,426 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/.tmp/hbase.id]:[hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/hbase.id] 2024-12-02T15:44:27,444 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T15:44:27,444 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T15:44:27,446 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-12-02T15:44:27,677 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,677 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,677 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,677 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741827_1003 (size=196) 2024-12-02T15:44:27,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741827_1003 (size=196) 2024-12-02T15:44:27,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741827_1003 (size=196) 2024-12-02T15:44:27,689 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T15:44:27,690 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T15:44:27,690 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T15:44:27,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741828_1004 (size=1189) 2024-12-02T15:44:27,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741828_1004 (size=1189) 2024-12-02T15:44:27,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741828_1004 (size=1189) 2024-12-02T15:44:27,707 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store 2024-12-02T15:44:27,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741829_1005 (size=34) 2024-12-02T15:44:27,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741829_1005 (size=34) 2024-12-02T15:44:27,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741829_1005 (size=34) 2024-12-02T15:44:27,722 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:27,723 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T15:44:27,723 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:27,723 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:27,723 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T15:44:27,723 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:27,723 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:27,723 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733154267723Disabling compacts and flushes for region at 1733154267723Disabling writes for close at 1733154267723Writing region close event to WAL at 1733154267723Closed at 1733154267723 2024-12-02T15:44:27,725 WARN [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/.initializing 2024-12-02T15:44:27,725 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/WALs/e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:27,730 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C42329%2C1733154266842, suffix=, logDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/WALs/e1c0b6c2739c,42329,1733154266842, archiveDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/oldWALs, maxLogs=10 2024-12-02T15:44:27,731 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor e1c0b6c2739c%2C42329%2C1733154266842.1733154267730 2024-12-02T15:44:27,744 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/WALs/e1c0b6c2739c,42329,1733154266842/e1c0b6c2739c%2C42329%2C1733154266842.1733154267730 2024-12-02T15:44:27,750 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43789:43789),(127.0.0.1/127.0.0.1:44569:44569),(127.0.0.1/127.0.0.1:46117:46117)] 2024-12-02T15:44:27,751 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T15:44:27,751 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:27,751 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,751 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,753 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,755 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T15:44:27,755 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:27,756 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:27,756 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,758 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T15:44:27,758 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:27,759 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T15:44:27,759 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,761 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T15:44:27,761 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:27,762 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T15:44:27,762 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,764 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T15:44:27,764 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:27,765 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T15:44:27,765 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,766 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,766 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,767 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,767 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,768 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T15:44:27,769 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T15:44:27,771 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T15:44:27,772 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74876601, jitterRate=0.11574830114841461}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T15:44:27,773 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733154267751Initializing all the Stores at 1733154267753 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154267753Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154267753Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154267753Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154267753Cleaning up temporary data from old regions at 1733154267767 (+14 ms)Region opened successfully at 1733154267772 (+5 ms) 2024-12-02T15:44:27,773 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T15:44:27,777 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7d69b182, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=e1c0b6c2739c/172.17.0.3:0 2024-12-02T15:44:27,778 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T15:44:27,778 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T15:44:27,778 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T15:44:27,778 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T15:44:27,779 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T15:44:27,779 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-02T15:44:27,779 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T15:44:27,781 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T15:44:27,782 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T15:44:27,792 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T15:44:27,793 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T15:44:27,794 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T15:44:27,803 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T15:44:27,803 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T15:44:27,805 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T15:44:27,811 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T15:44:27,812 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T15:44:27,819 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T15:44:27,822 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T15:44:27,884 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T15:44:27,895 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:27,895 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:27,895 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:27,895 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,895 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:27,895 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,895 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,895 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,896 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=e1c0b6c2739c,42329,1733154266842, sessionid=0x101980101330000, setting cluster-up flag (Was=false) 2024-12-02T15:44:27,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,911 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,911 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,945 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T15:44:27,946 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:27,968 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,968 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,968 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,968 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:27,994 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T15:44:27,996 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:27,997 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T15:44:28,000 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T15:44:28,000 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T15:44:28,000 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T15:44:28,001 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: e1c0b6c2739c,42329,1733154266842 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T15:44:28,002 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/e1c0b6c2739c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T15:44:28,002 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/e1c0b6c2739c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T15:44:28,002 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T15:44:28,002 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T15:44:28,002 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/e1c0b6c2739c:0, corePoolSize=10, maxPoolSize=10 2024-12-02T15:44:28,003 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,003 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T15:44:28,003 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,003 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733154298003 2024-12-02T15:44:28,004 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T15:44:28,004 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T15:44:28,004 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T15:44:28,004 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T15:44:28,004 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T15:44:28,004 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T15:44:28,004 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,005 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T15:44:28,005 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T15:44:28,005 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T15:44:28,005 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T15:44:28,005 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T15:44:28,006 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T15:44:28,006 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T15:44:28,006 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.large.0-1733154268006,5,FailOnTimeoutGroup] 2024-12-02T15:44:28,006 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.small.0-1733154268006,5,FailOnTimeoutGroup] 2024-12-02T15:44:28,006 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,006 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T15:44:28,006 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,007 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,007 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,007 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T15:44:28,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741831_1007 (size=1321) 2024-12-02T15:44:28,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741831_1007 (size=1321) 2024-12-02T15:44:28,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741831_1007 (size=1321) 2024-12-02T15:44:28,018 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T15:44:28,018 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4 2024-12-02T15:44:28,026 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741832_1008 (size=32) 2024-12-02T15:44:28,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741832_1008 (size=32) 2024-12-02T15:44:28,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741832_1008 (size=32) 2024-12-02T15:44:28,028 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:28,029 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T15:44:28,031 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T15:44:28,031 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,032 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:28,032 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T15:44:28,034 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T15:44:28,034 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,035 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:28,035 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T15:44:28,037 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T15:44:28,037 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,038 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:28,038 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T15:44:28,040 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T15:44:28,040 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,041 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:28,041 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T15:44:28,042 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740 2024-12-02T15:44:28,042 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740 2024-12-02T15:44:28,044 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T15:44:28,044 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T15:44:28,045 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T15:44:28,046 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T15:44:28,051 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(746): ClusterId : f68683c7-d3bb-4f82-b24f-82d1f72fbf63 2024-12-02T15:44:28,051 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T15:44:28,051 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(746): ClusterId : f68683c7-d3bb-4f82-b24f-82d1f72fbf63 2024-12-02T15:44:28,051 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T15:44:28,051 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(746): ClusterId : f68683c7-d3bb-4f82-b24f-82d1f72fbf63 2024-12-02T15:44:28,051 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T15:44:28,054 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T15:44:28,056 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72495250, jitterRate=0.0802634060382843}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T15:44:28,057 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733154268028Initializing all the Stores at 1733154268029 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154268029Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154268029Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154268029Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154268029Cleaning up temporary data from old regions at 1733154268044 (+15 ms)Region opened successfully at 1733154268057 (+13 ms) 2024-12-02T15:44:28,057 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T15:44:28,057 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T15:44:28,057 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T15:44:28,057 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T15:44:28,057 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T15:44:28,061 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T15:44:28,062 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733154268057Disabling compacts and flushes for region at 1733154268057Disabling writes for close at 1733154268057Writing region close event to WAL at 1733154268061 (+4 ms)Closed at 1733154268061 2024-12-02T15:44:28,062 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T15:44:28,062 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T15:44:28,062 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T15:44:28,062 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T15:44:28,064 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T15:44:28,064 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T15:44:28,064 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T15:44:28,065 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T15:44:28,065 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T15:44:28,066 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T15:44:28,068 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T15:44:28,078 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T15:44:28,078 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T15:44:28,079 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T15:44:28,079 DEBUG [RS:0;e1c0b6c2739c:35335 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@63648111, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=e1c0b6c2739c/172.17.0.3:0 2024-12-02T15:44:28,079 DEBUG [RS:2;e1c0b6c2739c:45273 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@365050ab, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=e1c0b6c2739c/172.17.0.3:0 2024-12-02T15:44:28,079 DEBUG [RS:1;e1c0b6c2739c:36719 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@75c0db74, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=e1c0b6c2739c/172.17.0.3:0 2024-12-02T15:44:28,091 DEBUG [RS:2;e1c0b6c2739c:45273 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;e1c0b6c2739c:45273 2024-12-02T15:44:28,091 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T15:44:28,091 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T15:44:28,091 DEBUG [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T15:44:28,092 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,42329,1733154266842 with port=45273, startcode=1733154267086 2024-12-02T15:44:28,092 DEBUG [RS:2;e1c0b6c2739c:45273 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T15:44:28,094 DEBUG [RS:0;e1c0b6c2739c:35335 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;e1c0b6c2739c:35335 2024-12-02T15:44:28,094 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T15:44:28,095 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T15:44:28,095 DEBUG [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T15:44:28,096 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,42329,1733154266842 with port=35335, startcode=1733154266991 2024-12-02T15:44:28,096 DEBUG [RS:1;e1c0b6c2739c:36719 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;e1c0b6c2739c:36719 2024-12-02T15:44:28,096 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T15:44:28,096 DEBUG [RS:0;e1c0b6c2739c:35335 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T15:44:28,096 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T15:44:28,096 DEBUG [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T15:44:28,097 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(2659): reportForDuty to master=e1c0b6c2739c,42329,1733154266842 with port=36719, startcode=1733154267039 2024-12-02T15:44:28,097 DEBUG [RS:1;e1c0b6c2739c:36719 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T15:44:28,102 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:33977, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T15:44:28,102 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:52231, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T15:44:28,103 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42329 {}] master.ServerManager(363): Checking decommissioned status of RegionServer e1c0b6c2739c,45273,1733154267086 2024-12-02T15:44:28,103 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42329 {}] master.ServerManager(517): Registering regionserver=e1c0b6c2739c,45273,1733154267086 2024-12-02T15:44:28,103 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38521, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T15:44:28,105 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42329 {}] master.ServerManager(363): Checking decommissioned status of RegionServer e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,105 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42329 {}] master.ServerManager(517): Registering regionserver=e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,105 DEBUG [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4 2024-12-02T15:44:28,105 DEBUG [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41279 2024-12-02T15:44:28,106 DEBUG [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T15:44:28,108 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42329 {}] master.ServerManager(363): Checking decommissioned status of RegionServer e1c0b6c2739c,35335,1733154266991 2024-12-02T15:44:28,108 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42329 {}] master.ServerManager(517): Registering regionserver=e1c0b6c2739c,35335,1733154266991 2024-12-02T15:44:28,108 DEBUG [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4 2024-12-02T15:44:28,108 DEBUG [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41279 2024-12-02T15:44:28,108 DEBUG [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T15:44:28,111 DEBUG [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4 2024-12-02T15:44:28,111 DEBUG [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41279 2024-12-02T15:44:28,111 DEBUG [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T15:44:28,111 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T15:44:28,153 DEBUG [RS:2;e1c0b6c2739c:45273 {}] zookeeper.ZKUtil(111): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/e1c0b6c2739c,45273,1733154267086 2024-12-02T15:44:28,153 DEBUG [RS:1;e1c0b6c2739c:36719 {}] zookeeper.ZKUtil(111): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,153 WARN [RS:1;e1c0b6c2739c:36719 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T15:44:28,153 WARN [RS:2;e1c0b6c2739c:45273 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T15:44:28,153 INFO [RS:1;e1c0b6c2739c:36719 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T15:44:28,153 INFO [RS:2;e1c0b6c2739c:45273 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T15:44:28,153 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [e1c0b6c2739c,36719,1733154267039] 2024-12-02T15:44:28,153 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [e1c0b6c2739c,45273,1733154267086] 2024-12-02T15:44:28,153 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [e1c0b6c2739c,35335,1733154266991] 2024-12-02T15:44:28,153 DEBUG [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,153 DEBUG [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,45273,1733154267086 2024-12-02T15:44:28,154 DEBUG [RS:0;e1c0b6c2739c:35335 {}] zookeeper.ZKUtil(111): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/e1c0b6c2739c,35335,1733154266991 2024-12-02T15:44:28,154 WARN [RS:0;e1c0b6c2739c:35335 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T15:44:28,154 INFO [RS:0;e1c0b6c2739c:35335 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T15:44:28,154 DEBUG [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,35335,1733154266991 2024-12-02T15:44:28,159 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T15:44:28,162 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T15:44:28,162 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T15:44:28,162 INFO [RS:1;e1c0b6c2739c:36719 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T15:44:28,162 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T15:44:28,162 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,179 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T15:44:28,179 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T15:44:28,179 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T15:44:28,180 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T15:44:28,180 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,181 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,181 INFO [RS:2;e1c0b6c2739c:45273 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T15:44:28,181 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,181 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,181 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,181 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,181 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:28,182 DEBUG [RS:1;e1c0b6c2739c:36719 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:28,187 INFO [RS:0;e1c0b6c2739c:35335 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T15:44:28,187 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,191 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T15:44:28,192 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T15:44:28,193 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,193 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T15:44:28,193 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T15:44:28,193 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,193 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,193 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,193 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,193 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,193 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,193 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,36719,1733154267039-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T15:44:28,193 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,193 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/e1c0b6c2739c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:28,194 DEBUG [RS:2;e1c0b6c2739c:45273 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:28,194 DEBUG [RS:0;e1c0b6c2739c:35335 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T15:44:28,200 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,200 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,200 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,200 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,201 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,201 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,45273,1733154267086-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T15:44:28,200 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,201 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,201 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,201 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,201 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,202 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,35335,1733154266991-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T15:44:28,213 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T15:44:28,213 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,36719,1733154267039-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,214 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,214 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.Replication(171): e1c0b6c2739c,36719,1733154267039 started 2024-12-02T15:44:28,218 WARN [e1c0b6c2739c:42329 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-02T15:44:28,221 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T15:44:28,221 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,45273,1733154267086-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,222 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,222 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.Replication(171): e1c0b6c2739c,45273,1733154267086 started 2024-12-02T15:44:28,224 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T15:44:28,224 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,35335,1733154266991-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,224 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,225 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.Replication(171): e1c0b6c2739c,35335,1733154266991 started 2024-12-02T15:44:28,234 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,234 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1482): Serving as e1c0b6c2739c,36719,1733154267039, RpcServer on e1c0b6c2739c/172.17.0.3:36719, sessionid=0x101980101330002 2024-12-02T15:44:28,234 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T15:44:28,235 DEBUG [RS:1;e1c0b6c2739c:36719 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,235 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,36719,1733154267039' 2024-12-02T15:44:28,235 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T15:44:28,236 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T15:44:28,236 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T15:44:28,237 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T15:44:28,237 DEBUG [RS:1;e1c0b6c2739c:36719 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,237 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,36719,1733154267039' 2024-12-02T15:44:28,237 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T15:44:28,238 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T15:44:28,240 DEBUG [RS:1;e1c0b6c2739c:36719 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T15:44:28,240 INFO [RS:1;e1c0b6c2739c:36719 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T15:44:28,240 INFO [RS:1;e1c0b6c2739c:36719 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T15:44:28,243 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,244 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(1482): Serving as e1c0b6c2739c,45273,1733154267086, RpcServer on e1c0b6c2739c/172.17.0.3:45273, sessionid=0x101980101330003 2024-12-02T15:44:28,244 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T15:44:28,244 DEBUG [RS:2;e1c0b6c2739c:45273 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager e1c0b6c2739c,45273,1733154267086 2024-12-02T15:44:28,244 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,45273,1733154267086' 2024-12-02T15:44:28,244 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T15:44:28,245 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,245 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(1482): Serving as e1c0b6c2739c,35335,1733154266991, RpcServer on e1c0b6c2739c/172.17.0.3:35335, sessionid=0x101980101330001 2024-12-02T15:44:28,245 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T15:44:28,245 DEBUG [RS:0;e1c0b6c2739c:35335 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager e1c0b6c2739c,35335,1733154266991 2024-12-02T15:44:28,246 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,35335,1733154266991' 2024-12-02T15:44:28,246 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T15:44:28,246 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T15:44:28,246 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T15:44:28,246 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T15:44:28,246 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T15:44:28,247 DEBUG [RS:2;e1c0b6c2739c:45273 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager e1c0b6c2739c,45273,1733154267086 2024-12-02T15:44:28,247 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,45273,1733154267086' 2024-12-02T15:44:28,247 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T15:44:28,247 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T15:44:28,247 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T15:44:28,247 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T15:44:28,247 DEBUG [RS:0;e1c0b6c2739c:35335 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager e1c0b6c2739c,35335,1733154266991 2024-12-02T15:44:28,247 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'e1c0b6c2739c,35335,1733154266991' 2024-12-02T15:44:28,247 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T15:44:28,248 DEBUG [RS:2;e1c0b6c2739c:45273 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T15:44:28,248 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T15:44:28,248 INFO [RS:2;e1c0b6c2739c:45273 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T15:44:28,248 INFO [RS:2;e1c0b6c2739c:45273 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T15:44:28,248 DEBUG [RS:0;e1c0b6c2739c:35335 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T15:44:28,248 INFO [RS:0;e1c0b6c2739c:35335 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T15:44:28,248 INFO [RS:0;e1c0b6c2739c:35335 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T15:44:28,346 INFO [RS:1;e1c0b6c2739c:36719 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C36719%2C1733154267039, suffix=, logDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,36719,1733154267039, archiveDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/oldWALs, maxLogs=32 2024-12-02T15:44:28,348 INFO [RS:1;e1c0b6c2739c:36719 {}] monitor.StreamSlowMonitor(122): New stream slow monitor e1c0b6c2739c%2C36719%2C1733154267039.1733154268348 2024-12-02T15:44:28,353 INFO [RS:2;e1c0b6c2739c:45273 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C45273%2C1733154267086, suffix=, logDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,45273,1733154267086, archiveDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/oldWALs, maxLogs=32 2024-12-02T15:44:28,353 INFO [RS:2;e1c0b6c2739c:45273 {}] monitor.StreamSlowMonitor(122): New stream slow monitor e1c0b6c2739c%2C45273%2C1733154267086.1733154268353 2024-12-02T15:44:28,354 INFO [RS:0;e1c0b6c2739c:35335 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C35335%2C1733154266991, suffix=, logDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,35335,1733154266991, archiveDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/oldWALs, maxLogs=32 2024-12-02T15:44:28,358 INFO [RS:0;e1c0b6c2739c:35335 {}] monitor.StreamSlowMonitor(122): New stream slow monitor e1c0b6c2739c%2C35335%2C1733154266991.1733154268358 2024-12-02T15:44:28,384 INFO [RS:1;e1c0b6c2739c:36719 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,36719,1733154267039/e1c0b6c2739c%2C36719%2C1733154267039.1733154268348 2024-12-02T15:44:28,387 INFO [RS:0;e1c0b6c2739c:35335 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,35335,1733154266991/e1c0b6c2739c%2C35335%2C1733154266991.1733154268358 2024-12-02T15:44:28,387 INFO [RS:2;e1c0b6c2739c:45273 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,45273,1733154267086/e1c0b6c2739c%2C45273%2C1733154267086.1733154268353 2024-12-02T15:44:28,392 DEBUG [RS:1;e1c0b6c2739c:36719 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46117:46117),(127.0.0.1/127.0.0.1:43789:43789),(127.0.0.1/127.0.0.1:44569:44569)] 2024-12-02T15:44:28,392 DEBUG [RS:0;e1c0b6c2739c:35335 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46117:46117),(127.0.0.1/127.0.0.1:43789:43789),(127.0.0.1/127.0.0.1:44569:44569)] 2024-12-02T15:44:28,395 DEBUG [RS:2;e1c0b6c2739c:45273 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46117:46117),(127.0.0.1/127.0.0.1:44569:44569),(127.0.0.1/127.0.0.1:43789:43789)] 2024-12-02T15:44:28,469 DEBUG [e1c0b6c2739c:42329 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-02T15:44:28,469 DEBUG [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(204): Hosts are {e1c0b6c2739c=0} racks are {/default-rack=0} 2024-12-02T15:44:28,473 DEBUG [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-02T15:44:28,473 DEBUG [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-02T15:44:28,473 DEBUG [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-02T15:44:28,473 DEBUG [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-02T15:44:28,473 DEBUG [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-02T15:44:28,473 DEBUG [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-02T15:44:28,473 INFO [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-02T15:44:28,473 INFO [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-02T15:44:28,473 INFO [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-02T15:44:28,473 DEBUG [e1c0b6c2739c:42329 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-02T15:44:28,474 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,476 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as e1c0b6c2739c,36719,1733154267039, state=OPENING 2024-12-02T15:44:28,534 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T15:44:28,544 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:28,544 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:28,545 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:28,545 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:28,545 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:28,545 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:28,546 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:28,546 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T15:44:28,546 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:28,546 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=e1c0b6c2739c,36719,1733154267039}] 2024-12-02T15:44:28,700 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T15:44:28,703 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42011, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T15:44:28,708 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T15:44:28,708 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T15:44:28,711 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=e1c0b6c2739c%2C36719%2C1733154267039.meta, suffix=.meta, logDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,36719,1733154267039, archiveDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/oldWALs, maxLogs=32 2024-12-02T15:44:28,711 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor e1c0b6c2739c%2C36719%2C1733154267039.meta.1733154268711.meta 2024-12-02T15:44:28,725 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/WALs/e1c0b6c2739c,36719,1733154267039/e1c0b6c2739c%2C36719%2C1733154267039.meta.1733154268711.meta 2024-12-02T15:44:28,731 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:44569:44569),(127.0.0.1/127.0.0.1:43789:43789),(127.0.0.1/127.0.0.1:46117:46117)] 2024-12-02T15:44:28,738 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T15:44:28,738 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T15:44:28,738 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T15:44:28,739 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T15:44:28,739 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T15:44:28,739 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:28,739 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T15:44:28,739 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T15:44:28,743 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T15:44:28,744 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T15:44:28,745 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,745 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:28,745 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T15:44:28,747 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T15:44:28,747 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,748 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:28,748 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T15:44:28,749 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T15:44:28,749 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,750 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:28,750 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T15:44:28,751 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T15:44:28,752 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,752 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T15:44:28,753 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T15:44:28,754 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740 2024-12-02T15:44:28,755 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740 2024-12-02T15:44:28,757 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T15:44:28,757 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T15:44:28,758 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T15:44:28,760 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T15:44:28,761 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=62205918, jitterRate=-0.07305958867073059}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T15:44:28,762 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T15:44:28,763 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733154268739Writing region info on filesystem at 1733154268739Initializing all the Stores at 1733154268741 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154268741Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154268742 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154268742Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733154268742Cleaning up temporary data from old regions at 1733154268757 (+15 ms)Running coprocessor post-open hooks at 1733154268762 (+5 ms)Region opened successfully at 1733154268762 2024-12-02T15:44:28,765 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733154268700 2024-12-02T15:44:28,768 DEBUG [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T15:44:28,769 INFO [RS_OPEN_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T15:44:28,770 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,772 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as e1c0b6c2739c,36719,1733154267039, state=OPEN 2024-12-02T15:44:28,784 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T15:44:28,784 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T15:44:28,784 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T15:44:28,784 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T15:44:28,785 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:28,785 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:28,785 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:28,785 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T15:44:28,785 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:28,791 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T15:44:28,791 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=e1c0b6c2739c,36719,1733154267039 in 239 msec 2024-12-02T15:44:28,795 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T15:44:28,795 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 727 msec 2024-12-02T15:44:28,797 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T15:44:28,797 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T15:44:28,799 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T15:44:28,799 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=e1c0b6c2739c,36719,1733154267039, seqNum=-1] 2024-12-02T15:44:28,799 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T15:44:28,801 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:40785, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T15:44:28,814 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 813 msec 2024-12-02T15:44:28,814 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733154268814, completionTime=-1 2024-12-02T15:44:28,814 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-02T15:44:28,814 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T15:44:28,817 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-02T15:44:28,817 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733154328817 2024-12-02T15:44:28,817 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733154388817 2024-12-02T15:44:28,817 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 3 msec 2024-12-02T15:44:28,818 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-02T15:44:28,822 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,42329,1733154266842-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,822 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,42329,1733154266842-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,822 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,42329,1733154266842-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,822 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-e1c0b6c2739c:42329, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,822 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,822 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,825 DEBUG [master/e1c0b6c2739c:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T15:44:28,827 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.606sec 2024-12-02T15:44:28,828 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T15:44:28,828 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T15:44:28,828 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T15:44:28,828 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T15:44:28,828 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T15:44:28,828 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,42329,1733154266842-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T15:44:28,828 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,42329,1733154266842-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T15:44:28,831 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T15:44:28,831 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T15:44:28,831 INFO [master/e1c0b6c2739c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=e1c0b6c2739c,42329,1733154266842-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T15:44:28,852 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@52a57433, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T15:44:28,852 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request e1c0b6c2739c,42329,-1 for getting cluster id 2024-12-02T15:44:28,852 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T15:44:28,853 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'f68683c7-d3bb-4f82-b24f-82d1f72fbf63' 2024-12-02T15:44:28,854 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T15:44:28,854 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "f68683c7-d3bb-4f82-b24f-82d1f72fbf63" 2024-12-02T15:44:28,854 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7a57e884, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T15:44:28,854 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [e1c0b6c2739c,42329,-1] 2024-12-02T15:44:28,855 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T15:44:28,855 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:28,857 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56154, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T15:44:28,858 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@438ad0b4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T15:44:28,859 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T15:44:28,860 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=e1c0b6c2739c,36719,1733154267039, seqNum=-1] 2024-12-02T15:44:28,861 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T15:44:28,868 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:53556, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T15:44:28,874 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:28,875 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-02T15:44:28,876 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:28,876 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@2b915cfe 2024-12-02T15:44:28,877 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T15:44:28,879 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56156, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T15:44:28,880 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T15:44:28,882 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-02T15:44:28,887 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T15:44:28,887 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:28,887 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-02T15:44:28,889 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T15:44:28,890 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T15:44:28,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741837_1013 (size=392) 2024-12-02T15:44:28,909 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741837_1013 (size=392) 2024-12-02T15:44:28,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741837_1013 (size=392) 2024-12-02T15:44:28,917 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => e55f16e4b7f33e79e81ce159677e4c9f, NAME => 'TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4 2024-12-02T15:44:28,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741838_1014 (size=51) 2024-12-02T15:44:28,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741838_1014 (size=51) 2024-12-02T15:44:28,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741838_1014 (size=51) 2024-12-02T15:44:28,938 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:28,938 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing e55f16e4b7f33e79e81ce159677e4c9f, disabling compactions & flushes 2024-12-02T15:44:28,938 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:28,938 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:28,938 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. after waiting 0 ms 2024-12-02T15:44:28,938 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:28,938 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:28,938 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for e55f16e4b7f33e79e81ce159677e4c9f: Waiting for close lock at 1733154268938Disabling compacts and flushes for region at 1733154268938Disabling writes for close at 1733154268938Writing region close event to WAL at 1733154268938Closed at 1733154268938 2024-12-02T15:44:28,940 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T15:44:28,940 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733154268940"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733154268940"}]},"ts":"1733154268940"} 2024-12-02T15:44:28,943 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-02T15:44:28,945 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T15:44:28,945 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733154268945"}]},"ts":"1733154268945"} 2024-12-02T15:44:28,948 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-02T15:44:28,948 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {e1c0b6c2739c=0} racks are {/default-rack=0} 2024-12-02T15:44:28,949 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-02T15:44:28,949 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-02T15:44:28,949 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-02T15:44:28,949 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-02T15:44:28,949 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-02T15:44:28,949 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-02T15:44:28,949 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-02T15:44:28,949 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-02T15:44:28,949 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-02T15:44:28,949 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-02T15:44:28,950 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=e55f16e4b7f33e79e81ce159677e4c9f, ASSIGN}] 2024-12-02T15:44:28,952 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=e55f16e4b7f33e79e81ce159677e4c9f, ASSIGN 2024-12-02T15:44:28,953 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=e55f16e4b7f33e79e81ce159677e4c9f, ASSIGN; state=OFFLINE, location=e1c0b6c2739c,36719,1733154267039; forceNewPlan=false, retain=false 2024-12-02T15:44:28,993 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T15:44:29,028 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T15:44:29,029 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-02T15:44:29,031 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T15:44:29,031 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-02T15:44:29,032 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-02T15:44:29,032 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-02T15:44:29,104 INFO [e1c0b6c2739c:42329 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-02T15:44:29,104 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=e55f16e4b7f33e79e81ce159677e4c9f, regionState=OPENING, regionLocation=e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:29,108 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=e55f16e4b7f33e79e81ce159677e4c9f, ASSIGN because future has completed 2024-12-02T15:44:29,109 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure e55f16e4b7f33e79e81ce159677e4c9f, server=e1c0b6c2739c,36719,1733154267039}] 2024-12-02T15:44:29,203 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T15:44:29,267 INFO [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,268 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => e55f16e4b7f33e79e81ce159677e4c9f, NAME => 'TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f.', STARTKEY => '', ENDKEY => ''} 2024-12-02T15:44:29,268 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,268 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T15:44:29,269 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,269 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,271 INFO [StoreOpener-e55f16e4b7f33e79e81ce159677e4c9f-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,273 INFO [StoreOpener-e55f16e4b7f33e79e81ce159677e4c9f-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region e55f16e4b7f33e79e81ce159677e4c9f columnFamilyName cf 2024-12-02T15:44:29,273 DEBUG [StoreOpener-e55f16e4b7f33e79e81ce159677e4c9f-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T15:44:29,274 INFO [StoreOpener-e55f16e4b7f33e79e81ce159677e4c9f-1 {}] regionserver.HStore(327): Store=e55f16e4b7f33e79e81ce159677e4c9f/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T15:44:29,274 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,274 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,275 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,275 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,275 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,277 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,279 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T15:44:29,280 INFO [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened e55f16e4b7f33e79e81ce159677e4c9f; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68414639, jitterRate=0.019457563757896423}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T15:44:29,280 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,280 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for e55f16e4b7f33e79e81ce159677e4c9f: Running coprocessor pre-open hook at 1733154269269Writing region info on filesystem at 1733154269269Initializing all the Stores at 1733154269270 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733154269270Cleaning up temporary data from old regions at 1733154269275 (+5 ms)Running coprocessor post-open hooks at 1733154269280 (+5 ms)Region opened successfully at 1733154269280 2024-12-02T15:44:29,282 INFO [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f., pid=6, masterSystemTime=1733154269262 2024-12-02T15:44:29,285 DEBUG [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,285 INFO [RS_OPEN_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,286 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=e55f16e4b7f33e79e81ce159677e4c9f, regionState=OPEN, openSeqNum=2, regionLocation=e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:29,289 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure e55f16e4b7f33e79e81ce159677e4c9f, server=e1c0b6c2739c,36719,1733154267039 because future has completed 2024-12-02T15:44:29,296 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T15:44:29,297 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure e55f16e4b7f33e79e81ce159677e4c9f, server=e1c0b6c2739c,36719,1733154267039 in 182 msec 2024-12-02T15:44:29,300 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T15:44:29,300 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=e55f16e4b7f33e79e81ce159677e4c9f, ASSIGN in 346 msec 2024-12-02T15:44:29,301 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T15:44:29,302 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733154269302"}]},"ts":"1733154269302"} 2024-12-02T15:44:29,307 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-02T15:44:29,309 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T15:44:29,312 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 429 msec 2024-12-02T15:44:29,514 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T15:44:29,515 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-02T15:44:29,515 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-02T15:44:29,515 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T15:44:29,520 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-02T15:44:29,520 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T15:44:29,520 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-02T15:44:29,527 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f., hostname=e1c0b6c2739c,36719,1733154267039, seqNum=2] 2024-12-02T15:44:29,534 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestHBaseWalOnEC 2024-12-02T15:44:29,537 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-02T15:44:29,539 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T15:44:29,541 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-02T15:44:29,545 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T15:44:29,545 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T15:44:29,643 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T15:44:29,700 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36719 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-02T15:44:29,700 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,701 INFO [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing e55f16e4b7f33e79e81ce159677e4c9f 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-02T15:44:29,722 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f/.tmp/cf/4d19626a637a42eba168e763c8c0212e is 36, key is row/cf:cq/1733154269530/Put/seqid=0 2024-12-02T15:44:29,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741839_1015 (size=4787) 2024-12-02T15:44:29,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741839_1015 (size=4787) 2024-12-02T15:44:29,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741839_1015 (size=4787) 2024-12-02T15:44:29,733 INFO [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f/.tmp/cf/4d19626a637a42eba168e763c8c0212e 2024-12-02T15:44:29,743 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f/.tmp/cf/4d19626a637a42eba168e763c8c0212e as hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f/cf/4d19626a637a42eba168e763c8c0212e 2024-12-02T15:44:29,752 INFO [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f/cf/4d19626a637a42eba168e763c8c0212e, entries=1, sequenceid=5, filesize=4.7 K 2024-12-02T15:44:29,754 INFO [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for e55f16e4b7f33e79e81ce159677e4c9f in 53ms, sequenceid=5, compaction requested=false 2024-12-02T15:44:29,754 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for e55f16e4b7f33e79e81ce159677e4c9f: 2024-12-02T15:44:29,754 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,754 DEBUG [RS_FLUSH_OPERATIONS-regionserver/e1c0b6c2739c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-02T15:44:29,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-02T15:44:29,760 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-02T15:44:29,760 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 212 msec 2024-12-02T15:44:29,764 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 227 msec 2024-12-02T15:44:29,853 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42329 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T15:44:29,854 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-02T15:44:29,858 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T15:44:29,858 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T15:44:29,858 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:29,858 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:29,858 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:29,858 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T15:44:29,859 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T15:44:29,859 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=784344886, stopped=false 2024-12-02T15:44:29,859 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=e1c0b6c2739c,42329,1733154266842 2024-12-02T15:44:29,926 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:29,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:29,926 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:29,926 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T15:44:29,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T15:44:29,926 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:29,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:29,926 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:29,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:29,927 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T15:44:29,927 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:29,927 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:29,927 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:29,927 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:29,927 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'e1c0b6c2739c,35335,1733154266991' ***** 2024-12-02T15:44:29,927 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T15:44:29,928 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'e1c0b6c2739c,36719,1733154267039' ***** 2024-12-02T15:44:29,928 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:29,928 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T15:44:29,928 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'e1c0b6c2739c,45273,1733154267086' ***** 2024-12-02T15:44:29,928 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T15:44:29,928 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T15:44:29,928 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T15:44:29,928 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T15:44:29,928 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T15:44:29,928 INFO [RS:0;e1c0b6c2739c:35335 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T15:44:29,928 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T15:44:29,928 INFO [RS:2;e1c0b6c2739c:45273 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T15:44:29,928 INFO [RS:0;e1c0b6c2739c:35335 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T15:44:29,928 INFO [RS:1;e1c0b6c2739c:36719 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T15:44:29,928 INFO [RS:2;e1c0b6c2739c:45273 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T15:44:29,928 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(959): stopping server e1c0b6c2739c,35335,1733154266991 2024-12-02T15:44:29,928 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(959): stopping server e1c0b6c2739c,45273,1733154267086 2024-12-02T15:44:29,928 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T15:44:29,928 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T15:44:29,928 INFO [RS:1;e1c0b6c2739c:36719 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T15:44:29,928 INFO [RS:0;e1c0b6c2739c:35335 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;e1c0b6c2739c:35335. 2024-12-02T15:44:29,928 INFO [RS:2;e1c0b6c2739c:45273 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;e1c0b6c2739c:45273. 2024-12-02T15:44:29,928 DEBUG [RS:2;e1c0b6c2739c:45273 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:29,928 DEBUG [RS:0;e1c0b6c2739c:35335 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:29,928 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(3091): Received CLOSE for e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,928 DEBUG [RS:2;e1c0b6c2739c:45273 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:29,928 DEBUG [RS:0;e1c0b6c2739c:35335 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:29,928 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T15:44:29,928 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(976): stopping server e1c0b6c2739c,45273,1733154267086; all regions closed. 2024-12-02T15:44:29,928 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(976): stopping server e1c0b6c2739c,35335,1733154266991; all regions closed. 2024-12-02T15:44:29,929 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T15:44:29,929 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(959): stopping server e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:29,929 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T15:44:29,929 INFO [RS:1;e1c0b6c2739c:36719 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;e1c0b6c2739c:36719. 2024-12-02T15:44:29,929 DEBUG [RS:1;e1c0b6c2739c:36719 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T15:44:29,929 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,929 DEBUG [RS:1;e1c0b6c2739c:36719 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:29,929 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T15:44:29,929 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T15:44:29,929 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,930 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T15:44:29,930 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T15:44:29,930 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing e55f16e4b7f33e79e81ce159677e4c9f, disabling compactions & flushes 2024-12-02T15:44:29,930 INFO [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,930 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,930 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,930 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. after waiting 0 ms 2024-12-02T15:44:29,930 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,930 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,930 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,930 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,930 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-02T15:44:29,930 DEBUG [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1325): Online Regions={e55f16e4b7f33e79e81ce159677e4c9f=TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f., 1588230740=hbase:meta,,1.1588230740} 2024-12-02T15:44:29,930 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,930 DEBUG [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, e55f16e4b7f33e79e81ce159677e4c9f 2024-12-02T15:44:29,930 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,931 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,931 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T15:44:29,931 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:29,931 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T15:44:29,931 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T15:44:29,931 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T15:44:29,931 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T15:44:29,931 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-02T15:44:29,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741834_1010 (size=93) 2024-12-02T15:44:29,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741834_1010 (size=93) 2024-12-02T15:44:29,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741834_1010 (size=93) 2024-12-02T15:44:29,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741835_1011 (size=93) 2024-12-02T15:44:29,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741835_1011 (size=93) 2024-12-02T15:44:29,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741835_1011 (size=93) 2024-12-02T15:44:29,942 DEBUG [RS:2;e1c0b6c2739c:45273 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/oldWALs 2024-12-02T15:44:29,942 INFO [RS:2;e1c0b6c2739c:45273 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog e1c0b6c2739c%2C45273%2C1733154267086:(num 1733154268353) 2024-12-02T15:44:29,942 DEBUG [RS:2;e1c0b6c2739c:45273 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:29,942 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:29,942 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T15:44:29,942 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/default/TestHBaseWalOnEC/e55f16e4b7f33e79e81ce159677e4c9f/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-02T15:44:29,943 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.ChoreService(370): Chore service for: regionserver/e1c0b6c2739c:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-02T15:44:29,943 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T15:44:29,943 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T15:44:29,943 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T15:44:29,943 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T15:44:29,943 INFO [RS:2;e1c0b6c2739c:45273 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:45273 2024-12-02T15:44:29,944 INFO [regionserver/e1c0b6c2739c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T15:44:29,945 INFO [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,945 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for e55f16e4b7f33e79e81ce159677e4c9f: Waiting for close lock at 1733154269930Running coprocessor pre-close hooks at 1733154269930Disabling compacts and flushes for region at 1733154269930Disabling writes for close at 1733154269930Writing region close event to WAL at 1733154269935 (+5 ms)Running coprocessor post-close hooks at 1733154269944 (+9 ms)Closed at 1733154269945 (+1 ms) 2024-12-02T15:44:29,945 DEBUG [RS_CLOSE_REGION-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f. 2024-12-02T15:44:29,946 DEBUG [RS:0;e1c0b6c2739c:35335 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/oldWALs 2024-12-02T15:44:29,946 INFO [RS:0;e1c0b6c2739c:35335 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog e1c0b6c2739c%2C35335%2C1733154266991:(num 1733154268358) 2024-12-02T15:44:29,946 DEBUG [RS:0;e1c0b6c2739c:35335 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:29,946 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:29,946 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T15:44:29,946 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.ChoreService(370): Chore service for: regionserver/e1c0b6c2739c:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-02T15:44:29,947 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T15:44:29,947 INFO [regionserver/e1c0b6c2739c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T15:44:29,947 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T15:44:29,947 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T15:44:29,947 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T15:44:29,947 INFO [RS:0;e1c0b6c2739c:35335 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:35335 2024-12-02T15:44:29,951 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/info/619cfd96fe554fb798b5831c4097c953 is 153, key is TestHBaseWalOnEC,,1733154268880.e55f16e4b7f33e79e81ce159677e4c9f./info:regioninfo/1733154269286/Put/seqid=0 2024-12-02T15:44:29,953 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/e1c0b6c2739c,45273,1733154267086 2024-12-02T15:44:29,953 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T15:44:29,953 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T15:44:29,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741840_1016 (size=6637) 2024-12-02T15:44:29,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741840_1016 (size=6637) 2024-12-02T15:44:29,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741840_1016 (size=6637) 2024-12-02T15:44:29,961 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/e1c0b6c2739c,35335,1733154266991 2024-12-02T15:44:29,961 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/info/619cfd96fe554fb798b5831c4097c953 2024-12-02T15:44:29,961 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [e1c0b6c2739c,45273,1733154267086] 2024-12-02T15:44:29,961 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T15:44:29,978 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/e1c0b6c2739c,45273,1733154267086 already deleted, retry=false 2024-12-02T15:44:29,978 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; e1c0b6c2739c,45273,1733154267086 expired; onlineServers=2 2024-12-02T15:44:29,978 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [e1c0b6c2739c,35335,1733154266991] 2024-12-02T15:44:29,986 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/e1c0b6c2739c,35335,1733154266991 already deleted, retry=false 2024-12-02T15:44:29,986 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; e1c0b6c2739c,35335,1733154266991 expired; onlineServers=1 2024-12-02T15:44:29,988 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/ns/11124a8cfc614ed0bc0fdc18e3466f51 is 43, key is default/ns:d/1733154268802/Put/seqid=0 2024-12-02T15:44:29,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741841_1017 (size=5153) 2024-12-02T15:44:29,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741841_1017 (size=5153) 2024-12-02T15:44:29,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741841_1017 (size=5153) 2024-12-02T15:44:29,996 INFO [regionserver/e1c0b6c2739c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:29,996 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/ns/11124a8cfc614ed0bc0fdc18e3466f51 2024-12-02T15:44:30,003 INFO [regionserver/e1c0b6c2739c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:30,018 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/table/7226b19a6b8949e4a096e49028e1680f is 52, key is TestHBaseWalOnEC/table:state/1733154269302/Put/seqid=0 2024-12-02T15:44:30,021 INFO [regionserver/e1c0b6c2739c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:30,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741842_1018 (size=5249) 2024-12-02T15:44:30,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741842_1018 (size=5249) 2024-12-02T15:44:30,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741842_1018 (size=5249) 2024-12-02T15:44:30,035 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/table/7226b19a6b8949e4a096e49028e1680f 2024-12-02T15:44:30,048 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/info/619cfd96fe554fb798b5831c4097c953 as hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/info/619cfd96fe554fb798b5831c4097c953 2024-12-02T15:44:30,059 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/info/619cfd96fe554fb798b5831c4097c953, entries=10, sequenceid=11, filesize=6.5 K 2024-12-02T15:44:30,061 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/ns/11124a8cfc614ed0bc0fdc18e3466f51 as hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/ns/11124a8cfc614ed0bc0fdc18e3466f51 2024-12-02T15:44:30,070 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:30,070 DEBUG [pool-336-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45273-0x101980101330003, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:30,073 INFO [RS:2;e1c0b6c2739c:45273 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T15:44:30,073 INFO [RS:2;e1c0b6c2739c:45273 {}] regionserver.HRegionServer(1031): Exiting; stopping=e1c0b6c2739c,45273,1733154267086; zookeeper connection closed. 2024-12-02T15:44:30,073 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/ns/11124a8cfc614ed0bc0fdc18e3466f51, entries=2, sequenceid=11, filesize=5.0 K 2024-12-02T15:44:30,076 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/.tmp/table/7226b19a6b8949e4a096e49028e1680f as hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/table/7226b19a6b8949e4a096e49028e1680f 2024-12-02T15:44:30,078 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:30,078 INFO [RS:0;e1c0b6c2739c:35335 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T15:44:30,078 DEBUG [pool-324-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35335-0x101980101330001, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:30,078 INFO [RS:0;e1c0b6c2739c:35335 {}] regionserver.HRegionServer(1031): Exiting; stopping=e1c0b6c2739c,35335,1733154266991; zookeeper connection closed. 2024-12-02T15:44:30,079 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@34421513 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@34421513 2024-12-02T15:44:30,080 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7bdc47ec {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7bdc47ec 2024-12-02T15:44:30,086 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/table/7226b19a6b8949e4a096e49028e1680f, entries=2, sequenceid=11, filesize=5.1 K 2024-12-02T15:44:30,088 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 157ms, sequenceid=11, compaction requested=false 2024-12-02T15:44:30,100 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-02T15:44:30,102 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T15:44:30,103 INFO [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T15:44:30,103 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733154269930Running coprocessor pre-close hooks at 1733154269931 (+1 ms)Disabling compacts and flushes for region at 1733154269931Disabling writes for close at 1733154269931Obtaining lock to block concurrent updates at 1733154269931Preparing flush snapshotting stores in 1588230740 at 1733154269931Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733154269931Flushing stores of hbase:meta,,1.1588230740 at 1733154269932 (+1 ms)Flushing 1588230740/info: creating writer at 1733154269932Flushing 1588230740/info: appending metadata at 1733154269951 (+19 ms)Flushing 1588230740/info: closing flushed file at 1733154269951Flushing 1588230740/ns: creating writer at 1733154269970 (+19 ms)Flushing 1588230740/ns: appending metadata at 1733154269988 (+18 ms)Flushing 1588230740/ns: closing flushed file at 1733154269988Flushing 1588230740/table: creating writer at 1733154270004 (+16 ms)Flushing 1588230740/table: appending metadata at 1733154270018 (+14 ms)Flushing 1588230740/table: closing flushed file at 1733154270018Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6daecd6: reopening flushed file at 1733154270046 (+28 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7b66ba5b: reopening flushed file at 1733154270059 (+13 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@356e1b88: reopening flushed file at 1733154270074 (+15 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 157ms, sequenceid=11, compaction requested=false at 1733154270088 (+14 ms)Writing region close event to WAL at 1733154270095 (+7 ms)Running coprocessor post-close hooks at 1733154270101 (+6 ms)Closed at 1733154270103 (+2 ms) 2024-12-02T15:44:30,103 DEBUG [RS_CLOSE_META-regionserver/e1c0b6c2739c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T15:44:30,131 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(976): stopping server e1c0b6c2739c,36719,1733154267039; all regions closed. 2024-12-02T15:44:30,131 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,131 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,131 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,131 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,132 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741836_1012 (size=2751) 2024-12-02T15:44:30,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741836_1012 (size=2751) 2024-12-02T15:44:30,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741836_1012 (size=2751) 2024-12-02T15:44:30,141 DEBUG [RS:1;e1c0b6c2739c:36719 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/oldWALs 2024-12-02T15:44:30,141 INFO [RS:1;e1c0b6c2739c:36719 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog e1c0b6c2739c%2C36719%2C1733154267039.meta:.meta(num 1733154268711) 2024-12-02T15:44:30,143 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,143 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,143 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,143 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,143 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741833_1009 (size=1298) 2024-12-02T15:44:30,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741833_1009 (size=1298) 2024-12-02T15:44:30,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741833_1009 (size=1298) 2024-12-02T15:44:30,150 DEBUG [RS:1;e1c0b6c2739c:36719 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/oldWALs 2024-12-02T15:44:30,150 INFO [RS:1;e1c0b6c2739c:36719 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog e1c0b6c2739c%2C36719%2C1733154267039:(num 1733154268348) 2024-12-02T15:44:30,150 DEBUG [RS:1;e1c0b6c2739c:36719 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T15:44:30,150 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T15:44:30,150 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T15:44:30,150 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.ChoreService(370): Chore service for: regionserver/e1c0b6c2739c:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-02T15:44:30,151 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T15:44:30,151 INFO [regionserver/e1c0b6c2739c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T15:44:30,151 INFO [RS:1;e1c0b6c2739c:36719 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:36719 2024-12-02T15:44:30,161 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T15:44:30,161 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/e1c0b6c2739c,36719,1733154267039 2024-12-02T15:44:30,161 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T15:44:30,161 ERROR [Time-limited test-EventThread {}] zookeeper.ClientCnxn$EventThread(581): Error while calling watcher. java.util.concurrent.RejectedExecutionException: Task org.apache.hadoop.hbase.trace.TraceUtil$$Lambda$377/0x00007f95248f60d8@5148e55 rejected from java.util.concurrent.ThreadPoolExecutor@12b1a2f[Shutting down, pool size = 1, active threads = 0, queued tasks = 0, completed tasks = 14] at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2065) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:833) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1360) ~[?:?] at java.util.concurrent.Executors$DelegatedExecutorService.execute(Executors.java:721) ~[?:?] at org.apache.hadoop.hbase.zookeeper.ZKWatcher.process(ZKWatcher.java:613) ~[hbase-zookeeper-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:579) ~[zookeeper-3.8.4.jar:3.8.4] at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:554) ~[zookeeper-3.8.4.jar:3.8.4] 2024-12-02T15:44:30,169 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [e1c0b6c2739c,36719,1733154267039] 2024-12-02T15:44:30,178 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/e1c0b6c2739c,36719,1733154267039 already deleted, retry=false 2024-12-02T15:44:30,178 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; e1c0b6c2739c,36719,1733154267039 expired; onlineServers=0 2024-12-02T15:44:30,178 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'e1c0b6c2739c,42329,1733154266842' ***** 2024-12-02T15:44:30,178 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T15:44:30,178 INFO [M:0;e1c0b6c2739c:42329 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T15:44:30,178 INFO [M:0;e1c0b6c2739c:42329 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T15:44:30,178 DEBUG [M:0;e1c0b6c2739c:42329 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T15:44:30,179 DEBUG [M:0;e1c0b6c2739c:42329 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T15:44:30,179 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T15:44:30,179 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.large.0-1733154268006 {}] cleaner.HFileCleaner(306): Exit Thread[master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.large.0-1733154268006,5,FailOnTimeoutGroup] 2024-12-02T15:44:30,179 INFO [M:0;e1c0b6c2739c:42329 {}] hbase.ChoreService(370): Chore service for: master/e1c0b6c2739c:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T15:44:30,179 INFO [M:0;e1c0b6c2739c:42329 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T15:44:30,179 DEBUG [master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.small.0-1733154268006 {}] cleaner.HFileCleaner(306): Exit Thread[master/e1c0b6c2739c:0:becomeActiveMaster-HFileCleaner.small.0-1733154268006,5,FailOnTimeoutGroup] 2024-12-02T15:44:30,179 DEBUG [M:0;e1c0b6c2739c:42329 {}] master.HMaster(1795): Stopping service threads 2024-12-02T15:44:30,179 INFO [M:0;e1c0b6c2739c:42329 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T15:44:30,179 INFO [M:0;e1c0b6c2739c:42329 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T15:44:30,180 INFO [M:0;e1c0b6c2739c:42329 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T15:44:30,180 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T15:44:30,186 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T15:44:30,186 DEBUG [M:0;e1c0b6c2739c:42329 {}] zookeeper.ZKUtil(347): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T15:44:30,186 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T15:44:30,186 WARN [M:0;e1c0b6c2739c:42329 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T15:44:30,187 INFO [M:0;e1c0b6c2739c:42329 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/.lastflushedseqids 2024-12-02T15:44:30,198 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741843_1019 (size=127) 2024-12-02T15:44:30,198 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741843_1019 (size=127) 2024-12-02T15:44:30,199 INFO [M:0;e1c0b6c2739c:42329 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T15:44:30,199 INFO [M:0;e1c0b6c2739c:42329 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T15:44:30,199 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T15:44:30,200 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:30,200 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:30,200 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T15:44:30,200 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:30,200 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.84 KB heapSize=34.13 KB 2024-12-02T15:44:30,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741843_1019 (size=127) 2024-12-02T15:44:30,223 DEBUG [M:0;e1c0b6c2739c:42329 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/4c56c1220f13470fb5078d21f7e06687 is 82, key is hbase:meta,,1/info:regioninfo/1733154268770/Put/seqid=0 2024-12-02T15:44:30,232 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741844_1020 (size=5672) 2024-12-02T15:44:30,234 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741844_1020 (size=5672) 2024-12-02T15:44:30,234 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/4c56c1220f13470fb5078d21f7e06687 2024-12-02T15:44:30,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741844_1020 (size=5672) 2024-12-02T15:44:30,267 DEBUG [M:0;e1c0b6c2739c:42329 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1be9ac7eb609459384de2291c3f735e4 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733154269311/Put/seqid=0 2024-12-02T15:44:30,270 INFO [RS:1;e1c0b6c2739c:36719 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T15:44:30,270 INFO [RS:1;e1c0b6c2739c:36719 {}] regionserver.HRegionServer(1031): Exiting; stopping=e1c0b6c2739c,36719,1733154267039; zookeeper connection closed. 2024-12-02T15:44:30,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:30,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36719-0x101980101330002, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:30,271 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@10f85f81 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@10f85f81 2024-12-02T15:44:30,271 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-02T15:44:30,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741845_1021 (size=6440) 2024-12-02T15:44:30,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741845_1021 (size=6440) 2024-12-02T15:44:30,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741845_1021 (size=6440) 2024-12-02T15:44:30,280 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.15 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1be9ac7eb609459384de2291c3f735e4 2024-12-02T15:44:30,307 DEBUG [M:0;e1c0b6c2739c:42329 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1eb7b26b7afc4f0b9fd8c04e16650b26 is 69, key is e1c0b6c2739c,35335,1733154266991/rs:state/1733154268108/Put/seqid=0 2024-12-02T15:44:30,316 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741846_1022 (size=5294) 2024-12-02T15:44:30,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741846_1022 (size=5294) 2024-12-02T15:44:30,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741846_1022 (size=5294) 2024-12-02T15:44:30,318 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1eb7b26b7afc4f0b9fd8c04e16650b26 2024-12-02T15:44:30,325 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/4c56c1220f13470fb5078d21f7e06687 as hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/4c56c1220f13470fb5078d21f7e06687 2024-12-02T15:44:30,334 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/4c56c1220f13470fb5078d21f7e06687, entries=8, sequenceid=72, filesize=5.5 K 2024-12-02T15:44:30,336 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1be9ac7eb609459384de2291c3f735e4 as hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/1be9ac7eb609459384de2291c3f735e4 2024-12-02T15:44:30,344 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/1be9ac7eb609459384de2291c3f735e4, entries=8, sequenceid=72, filesize=6.3 K 2024-12-02T15:44:30,345 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1eb7b26b7afc4f0b9fd8c04e16650b26 as hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/1eb7b26b7afc4f0b9fd8c04e16650b26 2024-12-02T15:44:30,352 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41279/user/jenkins/test-data/3176bf7d-ea19-8c97-4412-95b16dbc14a4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/1eb7b26b7afc4f0b9fd8c04e16650b26, entries=3, sequenceid=72, filesize=5.2 K 2024-12-02T15:44:30,354 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 154ms, sequenceid=72, compaction requested=false 2024-12-02T15:44:30,359 INFO [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T15:44:30,359 DEBUG [M:0;e1c0b6c2739c:42329 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733154270199Disabling compacts and flushes for region at 1733154270199Disabling writes for close at 1733154270200 (+1 ms)Obtaining lock to block concurrent updates at 1733154270200Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733154270200Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27480, getHeapSize=34880, getOffHeapSize=0, getCellsCount=85 at 1733154270200Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733154270201 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733154270202 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733154270223 (+21 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733154270223Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733154270243 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733154270267 (+24 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733154270267Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733154270286 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733154270306 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733154270306Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1e06ca7b: reopening flushed file at 1733154270324 (+18 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6afa54ed: reopening flushed file at 1733154270334 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@183ef9d7: reopening flushed file at 1733154270344 (+10 ms)Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 154ms, sequenceid=72, compaction requested=false at 1733154270354 (+10 ms)Writing region close event to WAL at 1733154270359 (+5 ms)Closed at 1733154270359 2024-12-02T15:44:30,360 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,360 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,360 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,360 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,360 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T15:44:30,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39273 is added to blk_1073741830_1006 (size=32683) 2024-12-02T15:44:30,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36501 is added to blk_1073741830_1006 (size=32683) 2024-12-02T15:44:30,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41631 is added to blk_1073741830_1006 (size=32683) 2024-12-02T15:44:30,365 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T15:44:30,365 INFO [M:0;e1c0b6c2739c:42329 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T15:44:30,366 INFO [M:0;e1c0b6c2739c:42329 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:42329 2024-12-02T15:44:30,366 INFO [M:0;e1c0b6c2739c:42329 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T15:44:30,476 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:30,476 INFO [M:0;e1c0b6c2739c:42329 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T15:44:30,476 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42329-0x101980101330000, quorum=127.0.0.1:62243, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T15:44:30,479 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@bff0a43{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:30,479 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@19dff04d{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T15:44:30,479 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T15:44:30,479 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@38da8210{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T15:44:30,479 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6e0095f0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir/,STOPPED} 2024-12-02T15:44:30,480 WARN [BP-788991653-172.17.0.3-1733154264965 heartbeating to localhost/127.0.0.1:41279 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T15:44:30,480 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T15:44:30,481 WARN [BP-788991653-172.17.0.3-1733154264965 heartbeating to localhost/127.0.0.1:41279 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-788991653-172.17.0.3-1733154264965 (Datanode Uuid 07fcc5c6-5f77-4614-994c-f6ea8a3d5991) service to localhost/127.0.0.1:41279 2024-12-02T15:44:30,481 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T15:44:30,481 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data5/current/BP-788991653-172.17.0.3-1733154264965 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:30,481 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data6/current/BP-788991653-172.17.0.3-1733154264965 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:30,482 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T15:44:30,487 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6e5e4927{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:30,488 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1768a8c1{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T15:44:30,488 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T15:44:30,488 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@bb1336{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T15:44:30,488 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5b4297c4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir/,STOPPED} 2024-12-02T15:44:30,489 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T15:44:30,489 WARN [BP-788991653-172.17.0.3-1733154264965 heartbeating to localhost/127.0.0.1:41279 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T15:44:30,490 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T15:44:30,490 WARN [BP-788991653-172.17.0.3-1733154264965 heartbeating to localhost/127.0.0.1:41279 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-788991653-172.17.0.3-1733154264965 (Datanode Uuid 9d116beb-8d2b-4bee-b153-cb6b098221bd) service to localhost/127.0.0.1:41279 2024-12-02T15:44:30,490 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data3/current/BP-788991653-172.17.0.3-1733154264965 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:30,491 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data4/current/BP-788991653-172.17.0.3-1733154264965 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:30,491 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T15:44:30,496 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@38e5384{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T15:44:30,496 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7d6118e0{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T15:44:30,496 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T15:44:30,496 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@61a92fea{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T15:44:30,496 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2807f8c2{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir/,STOPPED} 2024-12-02T15:44:30,497 WARN [BP-788991653-172.17.0.3-1733154264965 heartbeating to localhost/127.0.0.1:41279 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T15:44:30,497 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T15:44:30,497 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T15:44:30,497 WARN [BP-788991653-172.17.0.3-1733154264965 heartbeating to localhost/127.0.0.1:41279 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-788991653-172.17.0.3-1733154264965 (Datanode Uuid 55b82f5a-737b-47d4-a608-9da47531bbcb) service to localhost/127.0.0.1:41279 2024-12-02T15:44:30,498 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data1/current/BP-788991653-172.17.0.3-1733154264965 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:30,498 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/cluster_b5d24819-c49a-f9b7-6998-a68c71b965dd/data/data2/current/BP-788991653-172.17.0.3-1733154264965 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T15:44:30,499 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T15:44:30,505 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7aaeb6cf{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T15:44:30,506 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@686c9dd5{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T15:44:30,506 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T15:44:30,506 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@56aa9d3b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T15:44:30,506 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@a49b909{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/2fc66bab-f9ce-36f0-d2d0-313d839db0fa/hadoop.log.dir/,STOPPED} 2024-12-02T15:44:30,513 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T15:44:30,544 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T15:44:30,551 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=151 (was 91) - Thread LEAK? -, OpenFileDescriptor=518 (was 439) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=633 (was 636), ProcessCount=11 (was 11), AvailableMemoryMB=9156 (was 9607)