2024-12-09 10:50:44,948 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-09 10:50:44,962 main DEBUG Took 0.011174 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-09 10:50:44,962 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-09 10:50:44,963 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-09 10:50:44,964 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-09 10:50:44,965 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:44,974 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-09 10:50:44,993 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:44,995 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:44,995 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:44,996 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:44,996 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:44,996 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:44,997 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:44,997 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:44,998 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:44,998 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:44,999 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:44,999 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:44,999 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:45,000 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:45,000 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:45,000 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:45,001 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:45,001 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:45,001 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:45,001 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:45,002 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:45,002 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:45,002 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:45,003 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-09 10:50:45,003 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:45,003 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-09 10:50:45,005 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-09 10:50:45,006 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-09 10:50:45,008 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-09 10:50:45,009 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-09 10:50:45,010 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-09 10:50:45,010 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-09 10:50:45,018 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-09 10:50:45,022 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-09 10:50:45,024 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-09 10:50:45,024 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-09 10:50:45,024 main DEBUG createAppenders(={Console}) 2024-12-09 10:50:45,025 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-09 10:50:45,025 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-09 10:50:45,026 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-09 10:50:45,026 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-09 10:50:45,026 main DEBUG OutputStream closed 2024-12-09 10:50:45,027 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-09 10:50:45,027 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-09 10:50:45,027 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-09 10:50:45,098 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-09 10:50:45,100 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-09 10:50:45,101 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-09 10:50:45,102 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-09 10:50:45,103 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-09 10:50:45,103 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-09 10:50:45,103 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-09 10:50:45,104 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-09 10:50:45,104 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-09 10:50:45,104 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-09 10:50:45,105 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-09 10:50:45,105 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-09 10:50:45,105 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-09 10:50:45,105 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-09 10:50:45,106 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-09 10:50:45,106 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-09 10:50:45,106 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-09 10:50:45,107 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-09 10:50:45,109 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-09 10:50:45,110 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-09 10:50:45,110 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-09 10:50:45,111 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-09T10:50:45,127 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-09 10:50:45,130 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-09 10:50:45,130 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-09T10:50:45,383 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0 2024-12-09T10:50:45,409 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3, deleteOnExit=true 2024-12-09T10:50:45,410 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/test.cache.data in system properties and HBase conf 2024-12-09T10:50:45,410 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.tmp.dir in system properties and HBase conf 2024-12-09T10:50:45,411 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir in system properties and HBase conf 2024-12-09T10:50:45,411 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-09T10:50:45,412 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-09T10:50:45,412 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-09T10:50:45,510 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-09T10:50:45,604 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-09T10:50:45,608 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-09T10:50:45,608 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-09T10:50:45,609 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-09T10:50:45,609 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-09T10:50:45,610 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-09T10:50:45,610 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-09T10:50:45,611 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-09T10:50:45,611 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-09T10:50:45,611 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-09T10:50:45,612 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/nfs.dump.dir in system properties and HBase conf 2024-12-09T10:50:45,612 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/java.io.tmpdir in system properties and HBase conf 2024-12-09T10:50:45,612 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-09T10:50:45,613 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-09T10:50:45,613 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-09T10:50:46,510 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-09T10:50:46,596 INFO [Time-limited test {}] log.Log(170): Logging initialized @2366ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-09T10:50:46,689 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:46,770 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T10:50:46,798 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T10:50:46,799 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T10:50:46,800 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-09T10:50:46,815 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:46,818 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir/,AVAILABLE} 2024-12-09T10:50:46,819 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T10:50:47,028 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@76e4c45c{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/java.io.tmpdir/jetty-localhost-36623-hadoop-hdfs-3_4_1-tests_jar-_-any-3314186079494574374/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-09T10:50:47,037 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:36623} 2024-12-09T10:50:47,038 INFO [Time-limited test {}] server.Server(415): Started @2809ms 2024-12-09T10:50:47,441 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:47,451 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T10:50:47,452 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T10:50:47,452 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T10:50:47,452 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-09T10:50:47,453 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@550154bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir/,AVAILABLE} 2024-12-09T10:50:47,454 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1a2478ad{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T10:50:47,583 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4839957b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/java.io.tmpdir/jetty-localhost-40055-hadoop-hdfs-3_4_1-tests_jar-_-any-8398628554440339761/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:47,584 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5306f615{HTTP/1.1, (http/1.1)}{localhost:40055} 2024-12-09T10:50:47,584 INFO [Time-limited test {}] server.Server(415): Started @3355ms 2024-12-09T10:50:47,646 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T10:50:47,783 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:47,787 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T10:50:47,790 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T10:50:47,790 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T10:50:47,790 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-09T10:50:47,791 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6463ad04{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir/,AVAILABLE} 2024-12-09T10:50:47,792 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7fa8fa5c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T10:50:47,933 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1c6b8f01{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/java.io.tmpdir/jetty-localhost-46559-hadoop-hdfs-3_4_1-tests_jar-_-any-8155743878328108339/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:47,934 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11f28dd2{HTTP/1.1, (http/1.1)}{localhost:46559} 2024-12-09T10:50:47,934 INFO [Time-limited test {}] server.Server(415): Started @3706ms 2024-12-09T10:50:47,937 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T10:50:47,984 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:47,990 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T10:50:47,993 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T10:50:47,993 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T10:50:47,993 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-09T10:50:47,994 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@c62369b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir/,AVAILABLE} 2024-12-09T10:50:47,995 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@24f92c39{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T10:50:48,116 WARN [Thread-107 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data2/current/BP-1773582303-172.17.0.3-1733741446266/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:48,116 WARN [Thread-108 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data4/current/BP-1773582303-172.17.0.3-1733741446266/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:48,116 WARN [Thread-105 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data1/current/BP-1773582303-172.17.0.3-1733741446266/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:48,116 WARN [Thread-106 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data3/current/BP-1773582303-172.17.0.3-1733741446266/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:48,141 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2e59159d{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/java.io.tmpdir/jetty-localhost-46653-hadoop-hdfs-3_4_1-tests_jar-_-any-7592683019318178133/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:48,142 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@a8e922f{HTTP/1.1, (http/1.1)}{localhost:46653} 2024-12-09T10:50:48,142 INFO [Time-limited test {}] server.Server(415): Started @3913ms 2024-12-09T10:50:48,145 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T10:50:48,165 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T10:50:48,165 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T10:50:48,237 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xed5e3a242ebe3d9d with lease ID 0x5a967d5eb821768a: Processing first storage report for DS-c2bc9b9d-3754-4367-9e2e-e97769f1c6fd from datanode DatanodeRegistration(127.0.0.1:37913, datanodeUuid=fef88b2f-b468-4668-9d86-df610b6e1f0e, infoPort=35205, infoSecurePort=0, ipcPort=37667, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266) 2024-12-09T10:50:48,238 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xed5e3a242ebe3d9d with lease ID 0x5a967d5eb821768a: from storage DS-c2bc9b9d-3754-4367-9e2e-e97769f1c6fd node DatanodeRegistration(127.0.0.1:37913, datanodeUuid=fef88b2f-b468-4668-9d86-df610b6e1f0e, infoPort=35205, infoSecurePort=0, ipcPort=37667, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2024-12-09T10:50:48,239 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xae7465234323469f with lease ID 0x5a967d5eb821768b: Processing first storage report for DS-e41bba3d-9291-4448-ba5e-416685ed7833 from datanode DatanodeRegistration(127.0.0.1:38713, datanodeUuid=c8633e2b-43b2-4753-8101-90f583d169cb, infoPort=42101, infoSecurePort=0, ipcPort=34385, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266) 2024-12-09T10:50:48,239 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xae7465234323469f with lease ID 0x5a967d5eb821768b: from storage DS-e41bba3d-9291-4448-ba5e-416685ed7833 node DatanodeRegistration(127.0.0.1:38713, datanodeUuid=c8633e2b-43b2-4753-8101-90f583d169cb, infoPort=42101, infoSecurePort=0, ipcPort=34385, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-09T10:50:48,239 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xed5e3a242ebe3d9d with lease ID 0x5a967d5eb821768a: Processing first storage report for DS-8f8f6c4e-60bb-458f-8836-5631222de0b3 from datanode DatanodeRegistration(127.0.0.1:37913, datanodeUuid=fef88b2f-b468-4668-9d86-df610b6e1f0e, infoPort=35205, infoSecurePort=0, ipcPort=37667, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266) 2024-12-09T10:50:48,239 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xed5e3a242ebe3d9d with lease ID 0x5a967d5eb821768a: from storage DS-8f8f6c4e-60bb-458f-8836-5631222de0b3 node DatanodeRegistration(127.0.0.1:37913, datanodeUuid=fef88b2f-b468-4668-9d86-df610b6e1f0e, infoPort=35205, infoSecurePort=0, ipcPort=37667, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T10:50:48,239 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xae7465234323469f with lease ID 0x5a967d5eb821768b: Processing first storage report for DS-4b618197-3442-416b-8043-e417879541e9 from datanode DatanodeRegistration(127.0.0.1:38713, datanodeUuid=c8633e2b-43b2-4753-8101-90f583d169cb, infoPort=42101, infoSecurePort=0, ipcPort=34385, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266) 2024-12-09T10:50:48,240 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xae7465234323469f with lease ID 0x5a967d5eb821768b: from storage DS-4b618197-3442-416b-8043-e417879541e9 node DatanodeRegistration(127.0.0.1:38713, datanodeUuid=c8633e2b-43b2-4753-8101-90f583d169cb, infoPort=42101, infoSecurePort=0, ipcPort=34385, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T10:50:48,290 WARN [Thread-139 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data5/current/BP-1773582303-172.17.0.3-1733741446266/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:48,290 WARN [Thread-140 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data6/current/BP-1773582303-172.17.0.3-1733741446266/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:48,319 WARN [Thread-129 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T10:50:48,326 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd698ebda239eb7a0 with lease ID 0x5a967d5eb821768c: Processing first storage report for DS-e65cdba2-a3a6-4a27-b332-f0b08200a7d0 from datanode DatanodeRegistration(127.0.0.1:37003, datanodeUuid=b11aa68a-130a-4d14-902b-69cf740f7d6c, infoPort=43389, infoSecurePort=0, ipcPort=43939, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266) 2024-12-09T10:50:48,326 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd698ebda239eb7a0 with lease ID 0x5a967d5eb821768c: from storage DS-e65cdba2-a3a6-4a27-b332-f0b08200a7d0 node DatanodeRegistration(127.0.0.1:37003, datanodeUuid=b11aa68a-130a-4d14-902b-69cf740f7d6c, infoPort=43389, infoSecurePort=0, ipcPort=43939, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-09T10:50:48,326 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd698ebda239eb7a0 with lease ID 0x5a967d5eb821768c: Processing first storage report for DS-5c711505-6db1-4463-8dfa-b84dd14f446e from datanode DatanodeRegistration(127.0.0.1:37003, datanodeUuid=b11aa68a-130a-4d14-902b-69cf740f7d6c, infoPort=43389, infoSecurePort=0, ipcPort=43939, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266) 2024-12-09T10:50:48,327 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd698ebda239eb7a0 with lease ID 0x5a967d5eb821768c: from storage DS-5c711505-6db1-4463-8dfa-b84dd14f446e node DatanodeRegistration(127.0.0.1:37003, datanodeUuid=b11aa68a-130a-4d14-902b-69cf740f7d6c, infoPort=43389, infoSecurePort=0, ipcPort=43939, storageInfo=lv=-57;cid=testClusterID;nsid=974069069;c=1733741446266), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-09T10:50:48,549 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0 2024-12-09T10:50:48,628 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-09T10:50:48,702 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=159, OpenFileDescriptor=391, MaxFileDescriptor=1048576, SystemLoadAverage=198, ProcessCount=11, AvailableMemoryMB=8085 2024-12-09T10:50:48,704 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-09T10:50:48,726 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-09T10:50:48,821 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/zookeeper_0, clientPort=58213, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-09T10:50:48,833 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=58213 2024-12-09T10:50:48,867 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:48,871 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:48,978 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:48,978 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:49,029 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:49970 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:37003:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49970 dst: /127.0.0.1:37003 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:49,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-09T10:50:49,452 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:49,461 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db with version=8 2024-12-09T10:50:49,462 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/hbase-staging 2024-12-09T10:50:49,559 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-09T10:50:49,796 INFO [Time-limited test {}] client.ConnectionUtils(128): master/7ea43b39fc24:0 server-side Connection retries=45 2024-12-09T10:50:49,807 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:49,808 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:49,813 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T10:50:49,813 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:49,813 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T10:50:49,948 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-09T10:50:50,012 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-09T10:50:50,021 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-09T10:50:50,025 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T10:50:50,052 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 7954 (auto-detected) 2024-12-09T10:50:50,053 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:03 (auto-detected) 2024-12-09T10:50:50,071 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:40433 2024-12-09T10:50:50,093 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:40433 connecting to ZooKeeper ensemble=127.0.0.1:58213 2024-12-09T10:50:50,126 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:404330x0, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T10:50:50,128 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:40433-0x100bd5f2d610000 connected 2024-12-09T10:50:50,155 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,158 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,169 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:50,173 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db, hbase.cluster.distributed=false 2024-12-09T10:50:50,198 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T10:50:50,203 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40433 2024-12-09T10:50:50,207 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40433 2024-12-09T10:50:50,207 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40433 2024-12-09T10:50:50,211 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40433 2024-12-09T10:50:50,212 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40433 2024-12-09T10:50:50,333 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/7ea43b39fc24:0 server-side Connection retries=45 2024-12-09T10:50:50,334 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,335 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,335 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T10:50:50,335 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,335 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T10:50:50,338 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T10:50:50,341 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T10:50:50,342 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:36641 2024-12-09T10:50:50,345 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36641 connecting to ZooKeeper ensemble=127.0.0.1:58213 2024-12-09T10:50:50,346 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,350 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,358 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:366410x0, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T10:50:50,359 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:50,359 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36641-0x100bd5f2d610001 connected 2024-12-09T10:50:50,363 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T10:50:50,372 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T10:50:50,374 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T10:50:50,380 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T10:50:50,381 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36641 2024-12-09T10:50:50,381 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36641 2024-12-09T10:50:50,381 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36641 2024-12-09T10:50:50,383 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36641 2024-12-09T10:50:50,383 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36641 2024-12-09T10:50:50,403 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/7ea43b39fc24:0 server-side Connection retries=45 2024-12-09T10:50:50,403 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,403 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,404 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T10:50:50,404 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,404 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T10:50:50,404 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T10:50:50,405 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T10:50:50,406 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:46667 2024-12-09T10:50:50,408 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:46667 connecting to ZooKeeper ensemble=127.0.0.1:58213 2024-12-09T10:50:50,409 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,411 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,416 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:466670x0, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T10:50:50,417 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:46667-0x100bd5f2d610002 connected 2024-12-09T10:50:50,417 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:50,418 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T10:50:50,422 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T10:50:50,423 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T10:50:50,424 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T10:50:50,425 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46667 2024-12-09T10:50:50,425 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46667 2024-12-09T10:50:50,429 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46667 2024-12-09T10:50:50,430 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46667 2024-12-09T10:50:50,430 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46667 2024-12-09T10:50:50,448 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/7ea43b39fc24:0 server-side Connection retries=45 2024-12-09T10:50:50,448 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,448 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,449 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T10:50:50,449 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:50,449 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T10:50:50,449 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T10:50:50,449 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T10:50:50,450 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:33477 2024-12-09T10:50:50,451 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:33477 connecting to ZooKeeper ensemble=127.0.0.1:58213 2024-12-09T10:50:50,453 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,455 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,460 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:334770x0, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T10:50:50,461 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:334770x0, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:50,461 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:33477-0x100bd5f2d610003 connected 2024-12-09T10:50:50,462 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T10:50:50,463 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T10:50:50,464 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T10:50:50,466 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T10:50:50,467 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=33477 2024-12-09T10:50:50,467 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=33477 2024-12-09T10:50:50,467 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=33477 2024-12-09T10:50:50,468 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=33477 2024-12-09T10:50:50,468 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=33477 2024-12-09T10:50:50,489 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;7ea43b39fc24:40433 2024-12-09T10:50:50,491 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:50,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:50,498 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:50,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:50,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:50,500 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:50,527 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T10:50:50,527 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T10:50:50,527 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T10:50:50,527 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:50,527 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:50,527 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:50,527 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:50,528 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-09T10:50:50,530 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/7ea43b39fc24,40433,1733741449612 from backup master directory 2024-12-09T10:50:50,533 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:50,533 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:50,533 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:50,533 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:50,533 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:50,534 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T10:50:50,534 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:50,536 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-09T10:50:50,538 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-09T10:50:50,604 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/hbase.id] with ID: a45a09be-a3b8-4920-9f41-4f01d7b6701d 2024-12-09T10:50:50,604 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/.tmp/hbase.id 2024-12-09T10:50:50,611 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:50,611 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:50,614 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:54578 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:38713:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54578 dst: /127.0.0.1:38713 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:50,620 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-09T10:50:50,621 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:50,621 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/.tmp/hbase.id]:[hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/hbase.id] 2024-12-09T10:50:50,668 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:50,673 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-09T10:50:50,693 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 18ms. 2024-12-09T10:50:50,696 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:50,696 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:50,696 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:50,696 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:50,709 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:50,709 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:50,712 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:50006 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:37003:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50006 dst: /127.0.0.1:37003 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:50,718 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-09T10:50:50,719 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:50,735 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-09T10:50:50,737 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-09T10:50:50,743 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T10:50:50,773 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:50,773 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:50,778 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:54608 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:38713:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54608 dst: /127.0.0.1:38713 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:50,784 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-09T10:50:50,785 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:50,804 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store 2024-12-09T10:50:50,819 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:50,820 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:50,823 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:35464 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:37913:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35464 dst: /127.0.0.1:37913 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:50,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-09T10:50:50,831 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:50,836 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-09T10:50:50,840 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:50,841 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-09T10:50:50,841 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:50,841 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:50,843 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-09T10:50:50,843 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:50,843 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:50,844 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733741450841Disabling compacts and flushes for region at 1733741450841Disabling writes for close at 1733741450843 (+2 ms)Writing region close event to WAL at 1733741450843Closed at 1733741450843 2024-12-09T10:50:50,847 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/.initializing 2024-12-09T10:50:50,847 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/WALs/7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:50,856 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-09T10:50:50,872 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C40433%2C1733741449612, suffix=, logDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/WALs/7ea43b39fc24,40433,1733741449612, archiveDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/oldWALs, maxLogs=10 2024-12-09T10:50:50,905 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/WALs/7ea43b39fc24,40433,1733741449612/7ea43b39fc24%2C40433%2C1733741449612.1733741450878, exclude list is [], retry=0 2024-12-09T10:50:50,924 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:50,926 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38713,DS-e41bba3d-9291-4448-ba5e-416685ed7833,DISK] 2024-12-09T10:50:50,926 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37003,DS-e65cdba2-a3a6-4a27-b332-f0b08200a7d0,DISK] 2024-12-09T10:50:50,926 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37913,DS-c2bc9b9d-3754-4367-9e2e-e97769f1c6fd,DISK] 2024-12-09T10:50:50,929 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-09T10:50:50,973 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/WALs/7ea43b39fc24,40433,1733741449612/7ea43b39fc24%2C40433%2C1733741449612.1733741450878 2024-12-09T10:50:50,974 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:43389:43389),(127.0.0.1/127.0.0.1:35205:35205),(127.0.0.1/127.0.0.1:42101:42101)] 2024-12-09T10:50:50,974 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-09T10:50:50,975 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:50,978 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:50,979 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,019 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,046 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-09T10:50:51,050 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,052 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:51,053 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,056 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-09T10:50:51,056 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,057 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T10:50:51,057 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,060 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-09T10:50:51,060 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,061 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T10:50:51,061 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,064 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-09T10:50:51,064 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,065 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T10:50:51,065 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,069 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,071 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,078 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,078 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,082 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T10:50:51,086 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:51,092 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T10:50:51,093 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72318286, jitterRate=0.07762643694877625}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T10:50:51,099 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733741450992Initializing all the Stores at 1733741450994 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741450995 (+1 ms)Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741450996 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741450996Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741450996Cleaning up temporary data from old regions at 1733741451079 (+83 ms)Region opened successfully at 1733741451099 (+20 ms) 2024-12-09T10:50:51,100 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-09T10:50:51,135 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@bf4e8e0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=7ea43b39fc24/172.17.0.3:0 2024-12-09T10:50:51,167 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-09T10:50:51,179 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-09T10:50:51,179 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-09T10:50:51,182 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-09T10:50:51,184 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-09T10:50:51,190 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 6 msec 2024-12-09T10:50:51,190 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-09T10:50:51,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-09T10:50:51,217 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-09T10:50:51,218 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-09T10:50:51,227 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-09T10:50:51,229 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-09T10:50:51,232 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-09T10:50:51,233 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-09T10:50:51,236 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-09T10:50:51,238 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-09T10:50:51,242 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-09T10:50:51,244 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-09T10:50:51,245 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-09T10:50:51,247 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-09T10:50:51,265 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-09T10:50:51,267 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-09T10:50:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:51,272 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,272 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,275 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=7ea43b39fc24,40433,1733741449612, sessionid=0x100bd5f2d610000, setting cluster-up flag (Was=false) 2024-12-09T10:50:51,288 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,288 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,288 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,288 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,294 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-09T10:50:51,296 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:51,302 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,302 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,302 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,302 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:51,308 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-09T10:50:51,310 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:51,317 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-09T10:50:51,372 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(746): ClusterId : a45a09be-a3b8-4920-9f41-4f01d7b6701d 2024-12-09T10:50:51,372 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(746): ClusterId : a45a09be-a3b8-4920-9f41-4f01d7b6701d 2024-12-09T10:50:51,372 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(746): ClusterId : a45a09be-a3b8-4920-9f41-4f01d7b6701d 2024-12-09T10:50:51,376 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T10:50:51,376 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T10:50:51,376 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T10:50:51,381 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T10:50:51,381 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T10:50:51,381 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T10:50:51,382 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T10:50:51,382 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T10:50:51,382 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T10:50:51,386 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T10:50:51,386 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T10:50:51,386 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T10:50:51,387 DEBUG [RS:0;7ea43b39fc24:36641 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@37859844, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=7ea43b39fc24/172.17.0.3:0 2024-12-09T10:50:51,387 DEBUG [RS:2;7ea43b39fc24:33477 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@97cc50f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=7ea43b39fc24/172.17.0.3:0 2024-12-09T10:50:51,387 DEBUG [RS:1;7ea43b39fc24:46667 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@59167928, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=7ea43b39fc24/172.17.0.3:0 2024-12-09T10:50:51,391 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-09T10:50:51,400 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-09T10:50:51,407 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-09T10:50:51,414 DEBUG [RS:2;7ea43b39fc24:33477 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;7ea43b39fc24:33477 2024-12-09T10:50:51,414 DEBUG [RS:0;7ea43b39fc24:36641 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;7ea43b39fc24:36641 2024-12-09T10:50:51,414 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;7ea43b39fc24:46667 2024-12-09T10:50:51,412 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 7ea43b39fc24,40433,1733741449612 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-09T10:50:51,417 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T10:50:51,417 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T10:50:51,417 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T10:50:51,417 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T10:50:51,417 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T10:50:51,417 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T10:50:51,417 DEBUG [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T10:50:51,417 DEBUG [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T10:50:51,417 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T10:50:51,421 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(2659): reportForDuty to master=7ea43b39fc24,40433,1733741449612 with port=36641, startcode=1733741450293 2024-12-09T10:50:51,421 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(2659): reportForDuty to master=7ea43b39fc24,40433,1733741449612 with port=46667, startcode=1733741450402 2024-12-09T10:50:51,421 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(2659): reportForDuty to master=7ea43b39fc24,40433,1733741449612 with port=33477, startcode=1733741450447 2024-12-09T10:50:51,423 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/7ea43b39fc24:0, corePoolSize=5, maxPoolSize=5 2024-12-09T10:50:51,423 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/7ea43b39fc24:0, corePoolSize=5, maxPoolSize=5 2024-12-09T10:50:51,423 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=5, maxPoolSize=5 2024-12-09T10:50:51,423 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=5, maxPoolSize=5 2024-12-09T10:50:51,423 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/7ea43b39fc24:0, corePoolSize=10, maxPoolSize=10 2024-12-09T10:50:51,424 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,424 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=2, maxPoolSize=2 2024-12-09T10:50:51,424 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,429 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733741481429 2024-12-09T10:50:51,431 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T10:50:51,431 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-09T10:50:51,432 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-09T10:50:51,432 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-09T10:50:51,434 DEBUG [RS:2;7ea43b39fc24:33477 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T10:50:51,434 DEBUG [RS:0;7ea43b39fc24:36641 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T10:50:51,434 DEBUG [RS:1;7ea43b39fc24:46667 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T10:50:51,436 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-09T10:50:51,436 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-09T10:50:51,437 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-09T10:50:51,437 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-09T10:50:51,437 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,440 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,440 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-09T10:50:51,444 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-09T10:50:51,445 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-09T10:50:51,445 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-09T10:50:51,454 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-09T10:50:51,455 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-09T10:50:51,458 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:51,458 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.large.0-1733741451456,5,FailOnTimeoutGroup] 2024-12-09T10:50:51,458 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:51,459 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.small.0-1733741451458,5,FailOnTimeoutGroup] 2024-12-09T10:50:51,459 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,459 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-09T10:50:51,461 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,461 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,468 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:35484 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:37913:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35484 dst: /127.0.0.1:37913 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:51,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-09T10:50:51,485 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42231, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T10:50:51,485 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51725, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T10:50:51,485 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:47389, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T10:50:51,485 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:51,487 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-09T10:50:51,488 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db 2024-12-09T10:50:51,493 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40433 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:51,496 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40433 {}] master.ServerManager(517): Registering regionserver=7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:51,496 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:51,497 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:51,503 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:50050 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:37003:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50050 dst: /127.0.0.1:37003 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:51,509 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40433 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:51,509 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40433 {}] master.ServerManager(517): Registering regionserver=7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:51,510 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-09T10:50:51,511 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:51,512 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:51,513 DEBUG [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db 2024-12-09T10:50:51,513 DEBUG [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33501 2024-12-09T10:50:51,513 DEBUG [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T10:50:51,515 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40433 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 7ea43b39fc24,33477,1733741450447 2024-12-09T10:50:51,515 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db 2024-12-09T10:50:51,515 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40433 {}] master.ServerManager(517): Registering regionserver=7ea43b39fc24,33477,1733741450447 2024-12-09T10:50:51,515 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33501 2024-12-09T10:50:51,515 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T10:50:51,516 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-09T10:50:51,519 DEBUG [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db 2024-12-09T10:50:51,519 DEBUG [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33501 2024-12-09T10:50:51,519 DEBUG [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T10:50:51,520 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-09T10:50:51,520 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,521 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:51,522 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:51,522 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-09T10:50:51,522 DEBUG [RS:1;7ea43b39fc24:46667 {}] zookeeper.ZKUtil(111): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:51,522 DEBUG [RS:0;7ea43b39fc24:36641 {}] zookeeper.ZKUtil(111): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:51,522 WARN [RS:1;7ea43b39fc24:46667 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T10:50:51,522 WARN [RS:0;7ea43b39fc24:36641 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T10:50:51,522 INFO [RS:1;7ea43b39fc24:46667 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T10:50:51,523 INFO [RS:0;7ea43b39fc24:36641 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T10:50:51,523 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:51,523 DEBUG [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:51,524 DEBUG [RS:2;7ea43b39fc24:33477 {}] zookeeper.ZKUtil(111): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/7ea43b39fc24,33477,1733741450447 2024-12-09T10:50:51,524 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [7ea43b39fc24,36641,1733741450293] 2024-12-09T10:50:51,524 WARN [RS:2;7ea43b39fc24:33477 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T10:50:51,524 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [7ea43b39fc24,46667,1733741450402] 2024-12-09T10:50:51,524 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [7ea43b39fc24,33477,1733741450447] 2024-12-09T10:50:51,524 INFO [RS:2;7ea43b39fc24:33477 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T10:50:51,525 DEBUG [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,33477,1733741450447 2024-12-09T10:50:51,526 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-09T10:50:51,526 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,527 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:51,527 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-09T10:50:51,530 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-09T10:50:51,530 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,531 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:51,532 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-09T10:50:51,534 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-09T10:50:51,534 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:51,535 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:51,536 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-09T10:50:51,537 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740 2024-12-09T10:50:51,538 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740 2024-12-09T10:50:51,540 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-09T10:50:51,540 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-09T10:50:51,541 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T10:50:51,544 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-09T10:50:51,551 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T10:50:51,553 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60803672, jitterRate=-0.09395468235015869}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T10:50:51,555 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T10:50:51,555 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T10:50:51,555 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T10:50:51,556 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733741451513Initializing all the Stores at 1733741451515 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741451515Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741451515Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741451515Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741451516 (+1 ms)Cleaning up temporary data from old regions at 1733741451540 (+24 ms)Region opened successfully at 1733741451556 (+16 ms) 2024-12-09T10:50:51,556 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-09T10:50:51,556 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-09T10:50:51,556 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-09T10:50:51,556 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-09T10:50:51,556 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-09T10:50:51,558 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-09T10:50:51,558 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733741451556Disabling compacts and flushes for region at 1733741451556Disabling writes for close at 1733741451556Writing region close event to WAL at 1733741451557 (+1 ms)Closed at 1733741451557 2024-12-09T10:50:51,561 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T10:50:51,561 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-09T10:50:51,568 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-09T10:50:51,573 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T10:50:51,573 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T10:50:51,573 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T10:50:51,579 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-09T10:50:51,579 INFO [RS:2;7ea43b39fc24:33477 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T10:50:51,579 INFO [RS:0;7ea43b39fc24:36641 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T10:50:51,579 INFO [RS:1;7ea43b39fc24:46667 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T10:50:51,579 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,579 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,579 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,581 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T10:50:51,582 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T10:50:51,582 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T10:50:51,583 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-09T10:50:51,588 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T10:50:51,588 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T10:50:51,588 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T10:50:51,590 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,590 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,590 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,590 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,590 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/7ea43b39fc24:0, corePoolSize=2, maxPoolSize=2 2024-12-09T10:50:51,591 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/7ea43b39fc24:0, corePoolSize=2, maxPoolSize=2 2024-12-09T10:50:51,591 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/7ea43b39fc24:0, corePoolSize=2, maxPoolSize=2 2024-12-09T10:50:51,591 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,591 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:51,592 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,592 DEBUG [RS:0;7ea43b39fc24:36641 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:51,592 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,592 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,592 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:51,592 DEBUG [RS:2;7ea43b39fc24:33477 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:51,592 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,592 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,592 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,593 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,593 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,593 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:51,593 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:51,593 DEBUG [RS:1;7ea43b39fc24:46667 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:51,597 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,597 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,36641,1733741450293-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T10:50:51,598 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,46667,1733741450402-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T10:50:51,598 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,598 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,598 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,33477,1733741450447-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T10:50:51,620 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T10:50:51,620 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T10:50:51,620 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T10:50:51,622 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,36641,1733741450293-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,622 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,46667,1733741450402-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,622 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,33477,1733741450447-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,622 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,622 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,622 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,622 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.Replication(171): 7ea43b39fc24,46667,1733741450402 started 2024-12-09T10:50:51,623 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.Replication(171): 7ea43b39fc24,33477,1733741450447 started 2024-12-09T10:50:51,623 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.Replication(171): 7ea43b39fc24,36641,1733741450293 started 2024-12-09T10:50:51,642 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,642 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,642 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1482): Serving as 7ea43b39fc24,46667,1733741450402, RpcServer on 7ea43b39fc24/172.17.0.3:46667, sessionid=0x100bd5f2d610002 2024-12-09T10:50:51,642 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1482): Serving as 7ea43b39fc24,36641,1733741450293, RpcServer on 7ea43b39fc24/172.17.0.3:36641, sessionid=0x100bd5f2d610001 2024-12-09T10:50:51,643 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T10:50:51,643 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T10:50:51,643 DEBUG [RS:1;7ea43b39fc24:46667 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:51,643 DEBUG [RS:0;7ea43b39fc24:36641 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:51,643 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,46667,1733741450402' 2024-12-09T10:50:51,643 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,36641,1733741450293' 2024-12-09T10:50:51,643 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T10:50:51,643 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T10:50:51,645 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T10:50:51,645 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T10:50:51,645 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T10:50:51,645 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T10:50:51,645 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T10:50:51,645 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T10:50:51,646 DEBUG [RS:0;7ea43b39fc24:36641 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:51,646 DEBUG [RS:1;7ea43b39fc24:46667 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:51,646 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,36641,1733741450293' 2024-12-09T10:50:51,646 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,46667,1733741450402' 2024-12-09T10:50:51,646 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T10:50:51,646 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T10:50:51,646 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T10:50:51,646 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T10:50:51,647 DEBUG [RS:1;7ea43b39fc24:46667 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T10:50:51,647 DEBUG [RS:0;7ea43b39fc24:36641 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T10:50:51,647 INFO [RS:0;7ea43b39fc24:36641 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T10:50:51,647 INFO [RS:1;7ea43b39fc24:46667 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T10:50:51,647 INFO [RS:1;7ea43b39fc24:46667 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T10:50:51,647 INFO [RS:0;7ea43b39fc24:36641 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T10:50:51,648 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:51,648 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(1482): Serving as 7ea43b39fc24,33477,1733741450447, RpcServer on 7ea43b39fc24/172.17.0.3:33477, sessionid=0x100bd5f2d610003 2024-12-09T10:50:51,648 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T10:50:51,648 DEBUG [RS:2;7ea43b39fc24:33477 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 7ea43b39fc24,33477,1733741450447 2024-12-09T10:50:51,649 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,33477,1733741450447' 2024-12-09T10:50:51,649 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T10:50:51,649 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T10:50:51,650 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T10:50:51,650 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T10:50:51,650 DEBUG [RS:2;7ea43b39fc24:33477 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 7ea43b39fc24,33477,1733741450447 2024-12-09T10:50:51,650 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,33477,1733741450447' 2024-12-09T10:50:51,650 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T10:50:51,651 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T10:50:51,652 DEBUG [RS:2;7ea43b39fc24:33477 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T10:50:51,652 INFO [RS:2;7ea43b39fc24:33477 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T10:50:51,652 INFO [RS:2;7ea43b39fc24:33477 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T10:50:51,734 WARN [7ea43b39fc24:40433 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-09T10:50:51,753 INFO [RS:1;7ea43b39fc24:46667 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-09T10:50:51,753 INFO [RS:0;7ea43b39fc24:36641 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-09T10:50:51,753 INFO [RS:2;7ea43b39fc24:33477 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-09T10:50:51,756 INFO [RS:2;7ea43b39fc24:33477 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C33477%2C1733741450447, suffix=, logDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,33477,1733741450447, archiveDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/oldWALs, maxLogs=32 2024-12-09T10:50:51,756 INFO [RS:1;7ea43b39fc24:46667 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C46667%2C1733741450402, suffix=, logDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,46667,1733741450402, archiveDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/oldWALs, maxLogs=32 2024-12-09T10:50:51,756 INFO [RS:0;7ea43b39fc24:36641 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C36641%2C1733741450293, suffix=, logDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,36641,1733741450293, archiveDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/oldWALs, maxLogs=32 2024-12-09T10:50:51,778 DEBUG [RS:1;7ea43b39fc24:46667 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,46667,1733741450402/7ea43b39fc24%2C46667%2C1733741450402.1733741451763, exclude list is [], retry=0 2024-12-09T10:50:51,778 DEBUG [RS:2;7ea43b39fc24:33477 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,33477,1733741450447/7ea43b39fc24%2C33477%2C1733741450447.1733741451763, exclude list is [], retry=0 2024-12-09T10:50:51,778 DEBUG [RS:0;7ea43b39fc24:36641 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,36641,1733741450293/7ea43b39fc24%2C36641%2C1733741450293.1733741451763, exclude list is [], retry=0 2024-12-09T10:50:51,783 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38713,DS-e41bba3d-9291-4448-ba5e-416685ed7833,DISK] 2024-12-09T10:50:51,783 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38713,DS-e41bba3d-9291-4448-ba5e-416685ed7833,DISK] 2024-12-09T10:50:51,783 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37913,DS-c2bc9b9d-3754-4367-9e2e-e97769f1c6fd,DISK] 2024-12-09T10:50:51,784 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37913,DS-c2bc9b9d-3754-4367-9e2e-e97769f1c6fd,DISK] 2024-12-09T10:50:51,786 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37913,DS-c2bc9b9d-3754-4367-9e2e-e97769f1c6fd,DISK] 2024-12-09T10:50:51,786 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37003,DS-e65cdba2-a3a6-4a27-b332-f0b08200a7d0,DISK] 2024-12-09T10:50:51,811 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37003,DS-e65cdba2-a3a6-4a27-b332-f0b08200a7d0,DISK] 2024-12-09T10:50:51,811 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37003,DS-e65cdba2-a3a6-4a27-b332-f0b08200a7d0,DISK] 2024-12-09T10:50:51,812 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38713,DS-e41bba3d-9291-4448-ba5e-416685ed7833,DISK] 2024-12-09T10:50:51,814 INFO [RS:2;7ea43b39fc24:33477 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,33477,1733741450447/7ea43b39fc24%2C33477%2C1733741450447.1733741451763 2024-12-09T10:50:51,815 DEBUG [RS:2;7ea43b39fc24:33477 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:42101:42101),(127.0.0.1/127.0.0.1:35205:35205),(127.0.0.1/127.0.0.1:43389:43389)] 2024-12-09T10:50:51,818 INFO [RS:1;7ea43b39fc24:46667 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,46667,1733741450402/7ea43b39fc24%2C46667%2C1733741450402.1733741451763 2024-12-09T10:50:51,819 DEBUG [RS:1;7ea43b39fc24:46667 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:35205:35205),(127.0.0.1/127.0.0.1:42101:42101),(127.0.0.1/127.0.0.1:43389:43389)] 2024-12-09T10:50:51,820 INFO [RS:0;7ea43b39fc24:36641 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,36641,1733741450293/7ea43b39fc24%2C36641%2C1733741450293.1733741451763 2024-12-09T10:50:51,821 DEBUG [RS:0;7ea43b39fc24:36641 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:35205:35205),(127.0.0.1/127.0.0.1:43389:43389),(127.0.0.1/127.0.0.1:42101:42101)] 2024-12-09T10:50:51,986 DEBUG [7ea43b39fc24:40433 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-09T10:50:51,994 DEBUG [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(204): Hosts are {7ea43b39fc24=0} racks are {/default-rack=0} 2024-12-09T10:50:52,001 DEBUG [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-09T10:50:52,001 DEBUG [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-09T10:50:52,001 DEBUG [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-09T10:50:52,001 DEBUG [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-09T10:50:52,002 DEBUG [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-09T10:50:52,002 DEBUG [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-09T10:50:52,002 INFO [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-09T10:50:52,002 INFO [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-09T10:50:52,002 INFO [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-09T10:50:52,002 DEBUG [7ea43b39fc24:40433 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-09T10:50:52,009 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:52,017 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 7ea43b39fc24,46667,1733741450402, state=OPENING 2024-12-09T10:50:52,022 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-09T10:50:52,024 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:52,024 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:52,024 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:52,024 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:52,025 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:52,025 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:52,025 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:52,026 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:52,027 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-09T10:50:52,029 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=7ea43b39fc24,46667,1733741450402}] 2024-12-09T10:50:52,205 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-09T10:50:52,207 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37323, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-09T10:50:52,220 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-09T10:50:52,221 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-09T10:50:52,221 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-09T10:50:52,225 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C46667%2C1733741450402.meta, suffix=.meta, logDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,46667,1733741450402, archiveDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/oldWALs, maxLogs=32 2024-12-09T10:50:52,241 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,46667,1733741450402/7ea43b39fc24%2C46667%2C1733741450402.meta.1733741452226.meta, exclude list is [], retry=0 2024-12-09T10:50:52,246 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37003,DS-e65cdba2-a3a6-4a27-b332-f0b08200a7d0,DISK] 2024-12-09T10:50:52,246 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38713,DS-e41bba3d-9291-4448-ba5e-416685ed7833,DISK] 2024-12-09T10:50:52,247 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37913,DS-c2bc9b9d-3754-4367-9e2e-e97769f1c6fd,DISK] 2024-12-09T10:50:52,250 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/WALs/7ea43b39fc24,46667,1733741450402/7ea43b39fc24%2C46667%2C1733741450402.meta.1733741452226.meta 2024-12-09T10:50:52,251 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:43389:43389),(127.0.0.1/127.0.0.1:42101:42101),(127.0.0.1/127.0.0.1:35205:35205)] 2024-12-09T10:50:52,251 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-09T10:50:52,253 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-09T10:50:52,256 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-09T10:50:52,261 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-09T10:50:52,265 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-09T10:50:52,266 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:52,266 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-09T10:50:52,266 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-09T10:50:52,270 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-09T10:50:52,272 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-09T10:50:52,272 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:52,273 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:52,273 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-09T10:50:52,275 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-09T10:50:52,275 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:52,275 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:52,276 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-09T10:50:52,277 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-09T10:50:52,277 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:52,278 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:52,278 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-09T10:50:52,279 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-09T10:50:52,279 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:52,280 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:52,281 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-09T10:50:52,282 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740 2024-12-09T10:50:52,285 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740 2024-12-09T10:50:52,288 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-09T10:50:52,288 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-09T10:50:52,289 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T10:50:52,293 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-09T10:50:52,294 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63030706, jitterRate=-0.06076928973197937}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T10:50:52,295 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-09T10:50:52,297 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733741452267Writing region info on filesystem at 1733741452267Initializing all the Stores at 1733741452269 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741452269Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741452269Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741452269Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741452270 (+1 ms)Cleaning up temporary data from old regions at 1733741452288 (+18 ms)Running coprocessor post-open hooks at 1733741452295 (+7 ms)Region opened successfully at 1733741452297 (+2 ms) 2024-12-09T10:50:52,304 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733741452196 2024-12-09T10:50:52,316 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-09T10:50:52,317 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-09T10:50:52,318 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:52,320 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 7ea43b39fc24,46667,1733741450402, state=OPEN 2024-12-09T10:50:52,323 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T10:50:52,323 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T10:50:52,323 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T10:50:52,323 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T10:50:52,323 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:52,323 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:52,323 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:52,323 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:52,323 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:52,329 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-09T10:50:52,329 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=7ea43b39fc24,46667,1733741450402 in 296 msec 2024-12-09T10:50:52,335 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-09T10:50:52,335 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 763 msec 2024-12-09T10:50:52,337 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T10:50:52,337 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-09T10:50:52,359 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-09T10:50:52,360 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=7ea43b39fc24,46667,1733741450402, seqNum=-1] 2024-12-09T10:50:52,382 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T10:50:52,384 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:48323, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T10:50:52,434 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.0860 sec 2024-12-09T10:50:52,434 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733741452434, completionTime=-1 2024-12-09T10:50:52,439 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-09T10:50:52,439 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-09T10:50:52,472 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-09T10:50:52,472 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733741512472 2024-12-09T10:50:52,472 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733741572472 2024-12-09T10:50:52,472 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 32 msec 2024-12-09T10:50:52,474 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-09T10:50:52,481 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,40433,1733741449612-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:52,481 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,40433,1733741449612-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:52,481 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,40433,1733741449612-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:52,483 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-7ea43b39fc24:40433, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:52,484 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:52,484 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:52,490 DEBUG [master/7ea43b39fc24:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-09T10:50:52,513 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.979sec 2024-12-09T10:50:52,515 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-09T10:50:52,516 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-09T10:50:52,517 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-09T10:50:52,517 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-09T10:50:52,517 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-09T10:50:52,518 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,40433,1733741449612-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T10:50:52,519 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,40433,1733741449612-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-09T10:50:52,523 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-09T10:50:52,524 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-09T10:50:52,524 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,40433,1733741449612-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:52,586 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@65a66577, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T10:50:52,591 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-09T10:50:52,592 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-09T10:50:52,596 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 7ea43b39fc24,40433,-1 for getting cluster id 2024-12-09T10:50:52,600 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-09T10:50:52,609 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'a45a09be-a3b8-4920-9f41-4f01d7b6701d' 2024-12-09T10:50:52,612 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-09T10:50:52,613 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "a45a09be-a3b8-4920-9f41-4f01d7b6701d" 2024-12-09T10:50:52,615 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@22fce427, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T10:50:52,615 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [7ea43b39fc24,40433,-1] 2024-12-09T10:50:52,617 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-09T10:50:52,619 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:52,621 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42172, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-09T10:50:52,624 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6bf285b2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T10:50:52,624 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-09T10:50:52,633 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=7ea43b39fc24,46667,1733741450402, seqNum=-1] 2024-12-09T10:50:52,633 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T10:50:52,637 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35880, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T10:50:52,658 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:52,662 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-09T10:50:52,667 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:52,669 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@1669d569 2024-12-09T10:50:52,670 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-09T10:50:52,673 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42184, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-09T10:50:52,679 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-09T10:50:52,687 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-09T10:50:52,690 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-09T10:50:52,692 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-09T10:50:52,692 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:52,695 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-09T10:50:52,698 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:52,703 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:52,703 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:52,706 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:35546 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:37913:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35546 dst: /127.0.0.1:37913 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:52,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-09T10:50:52,712 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:52,714 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 465444a3d0faec8193e958662a333851, NAME => 'TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db 2024-12-09T10:50:52,720 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:52,720 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:52,725 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:35552 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:37913:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35552 dst: /127.0.0.1:37913 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:52,731 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-09T10:50:52,732 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:52,733 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:52,733 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 465444a3d0faec8193e958662a333851, disabling compactions & flushes 2024-12-09T10:50:52,733 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:52,733 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:52,733 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. after waiting 0 ms 2024-12-09T10:50:52,733 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:52,733 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:52,733 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 465444a3d0faec8193e958662a333851: Waiting for close lock at 1733741452733Disabling compacts and flushes for region at 1733741452733Disabling writes for close at 1733741452733Writing region close event to WAL at 1733741452733Closed at 1733741452733 2024-12-09T10:50:52,736 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-09T10:50:52,741 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733741452736"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733741452736"}]},"ts":"1733741452736"} 2024-12-09T10:50:52,746 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-09T10:50:52,748 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-09T10:50:52,751 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733741452748"}]},"ts":"1733741452748"} 2024-12-09T10:50:52,756 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-09T10:50:52,756 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {7ea43b39fc24=0} racks are {/default-rack=0} 2024-12-09T10:50:52,758 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-09T10:50:52,758 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-09T10:50:52,758 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-09T10:50:52,758 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-09T10:50:52,758 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-09T10:50:52,758 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-09T10:50:52,758 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-09T10:50:52,758 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-09T10:50:52,758 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-09T10:50:52,758 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-09T10:50:52,760 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=465444a3d0faec8193e958662a333851, ASSIGN}] 2024-12-09T10:50:52,762 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=465444a3d0faec8193e958662a333851, ASSIGN 2024-12-09T10:50:52,764 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=465444a3d0faec8193e958662a333851, ASSIGN; state=OFFLINE, location=7ea43b39fc24,36641,1733741450293; forceNewPlan=false, retain=false 2024-12-09T10:50:52,810 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:52,917 INFO [7ea43b39fc24:40433 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-09T10:50:52,918 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=465444a3d0faec8193e958662a333851, regionState=OPENING, regionLocation=7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:52,922 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=465444a3d0faec8193e958662a333851, ASSIGN because future has completed 2024-12-09T10:50:52,923 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 465444a3d0faec8193e958662a333851, server=7ea43b39fc24,36641,1733741450293}] 2024-12-09T10:50:53,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:53,078 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-09T10:50:53,080 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:39135, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-09T10:50:53,087 INFO [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:53,087 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 465444a3d0faec8193e958662a333851, NAME => 'TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851.', STARTKEY => '', ENDKEY => ''} 2024-12-09T10:50:53,088 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,088 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:53,088 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,088 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,090 INFO [StoreOpener-465444a3d0faec8193e958662a333851-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,093 INFO [StoreOpener-465444a3d0faec8193e958662a333851-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 465444a3d0faec8193e958662a333851 columnFamilyName cf 2024-12-09T10:50:53,093 DEBUG [StoreOpener-465444a3d0faec8193e958662a333851-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:53,094 INFO [StoreOpener-465444a3d0faec8193e958662a333851-1 {}] regionserver.HStore(327): Store=465444a3d0faec8193e958662a333851/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T10:50:53,094 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,095 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,096 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,097 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,097 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,099 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,104 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T10:50:53,105 INFO [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 465444a3d0faec8193e958662a333851; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70189203, jitterRate=0.045900627970695496}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-09T10:50:53,105 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 465444a3d0faec8193e958662a333851 2024-12-09T10:50:53,106 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 465444a3d0faec8193e958662a333851: Running coprocessor pre-open hook at 1733741453088Writing region info on filesystem at 1733741453088Initializing all the Stores at 1733741453090 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741453090Cleaning up temporary data from old regions at 1733741453097 (+7 ms)Running coprocessor post-open hooks at 1733741453105 (+8 ms)Region opened successfully at 1733741453106 (+1 ms) 2024-12-09T10:50:53,108 INFO [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851., pid=6, masterSystemTime=1733741453078 2024-12-09T10:50:53,112 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:53,112 INFO [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:53,113 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=465444a3d0faec8193e958662a333851, regionState=OPEN, openSeqNum=2, regionLocation=7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:53,118 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 465444a3d0faec8193e958662a333851, server=7ea43b39fc24,36641,1733741450293 because future has completed 2024-12-09T10:50:53,123 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-09T10:50:53,124 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 465444a3d0faec8193e958662a333851, server=7ea43b39fc24,36641,1733741450293 in 197 msec 2024-12-09T10:50:53,128 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-09T10:50:53,128 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=465444a3d0faec8193e958662a333851, ASSIGN in 364 msec 2024-12-09T10:50:53,129 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-09T10:50:53,129 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733741453129"}]},"ts":"1733741453129"} 2024-12-09T10:50:53,132 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-09T10:50:53,134 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-09T10:50:53,136 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 453 msec 2024-12-09T10:50:53,330 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:53,331 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-09T10:50:53,331 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-09T10:50:53,332 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-09T10:50:53,338 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-09T10:50:53,338 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-09T10:50:53,339 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-09T10:50:53,347 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851., hostname=7ea43b39fc24,36641,1733741450293, seqNum=2] 2024-12-09T10:50:53,349 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T10:50:53,351 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38082, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T10:50:53,361 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestHBaseWalOnEC 2024-12-09T10:50:53,366 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-09T10:50:53,369 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T10:50:53,369 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-09T10:50:53,371 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-09T10:50:53,373 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-09T10:50:53,480 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T10:50:53,536 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36641 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-09T10:50:53,537 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:53,541 INFO [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 465444a3d0faec8193e958662a333851 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-09T10:50:53,600 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851/.tmp/cf/4ee59fe9360240a28b020a0aa8b6b311 is 36, key is row/cf:cq/1733741453352/Put/seqid=0 2024-12-09T10:50:53,607 WARN [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:53,607 WARN [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:53,611 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1832355860_22 at /127.0.0.1:35564 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:37913:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35564 dst: /127.0.0.1:37913 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:53,615 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-09T10:50:53,616 WARN [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:53,616 INFO [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851/.tmp/cf/4ee59fe9360240a28b020a0aa8b6b311 2024-12-09T10:50:53,666 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851/.tmp/cf/4ee59fe9360240a28b020a0aa8b6b311 as hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851/cf/4ee59fe9360240a28b020a0aa8b6b311 2024-12-09T10:50:53,677 INFO [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851/cf/4ee59fe9360240a28b020a0aa8b6b311, entries=1, sequenceid=5, filesize=4.7 K 2024-12-09T10:50:53,684 INFO [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 465444a3d0faec8193e958662a333851 in 142ms, sequenceid=5, compaction requested=false 2024-12-09T10:50:53,685 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-09T10:50:53,687 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 465444a3d0faec8193e958662a333851: 2024-12-09T10:50:53,687 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:53,689 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-09T10:50:53,689 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T10:50:53,691 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-09T10:50:53,696 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-09T10:50:53,696 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 320 msec 2024-12-09T10:50:53,700 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 336 msec 2024-12-09T10:50:53,999 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40433 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T10:50:53,999 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-09T10:50:54,014 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-09T10:50:54,014 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-09T10:50:54,015 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:54,019 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,020 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,020 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-09T10:50:54,020 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-09T10:50:54,020 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=990235322, stopped=false 2024-12-09T10:50:54,021 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=7ea43b39fc24,40433,1733741449612 2024-12-09T10:50:54,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:54,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:54,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:54,023 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:54,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:54,023 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:54,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:54,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:54,023 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-09T10:50:54,024 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-09T10:50:54,024 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:54,024 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:54,024 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:54,024 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,024 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:54,025 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:54,025 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '7ea43b39fc24,36641,1733741450293' ***** 2024-12-09T10:50:54,025 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T10:50:54,025 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '7ea43b39fc24,46667,1733741450402' ***** 2024-12-09T10:50:54,025 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T10:50:54,025 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '7ea43b39fc24,33477,1733741450447' ***** 2024-12-09T10:50:54,025 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T10:50:54,025 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T10:50:54,026 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T10:50:54,026 INFO [RS:0;7ea43b39fc24:36641 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T10:50:54,026 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T10:50:54,026 INFO [RS:0;7ea43b39fc24:36641 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T10:50:54,026 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T10:50:54,026 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T10:50:54,026 INFO [RS:2;7ea43b39fc24:33477 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T10:50:54,026 INFO [RS:2;7ea43b39fc24:33477 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T10:50:54,026 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(3091): Received CLOSE for 465444a3d0faec8193e958662a333851 2024-12-09T10:50:54,026 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(959): stopping server 7ea43b39fc24,33477,1733741450447 2024-12-09T10:50:54,026 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T10:50:54,026 INFO [RS:2;7ea43b39fc24:33477 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;7ea43b39fc24:33477. 2024-12-09T10:50:54,026 DEBUG [RS:2;7ea43b39fc24:33477 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:54,026 DEBUG [RS:2;7ea43b39fc24:33477 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,027 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(976): stopping server 7ea43b39fc24,33477,1733741450447; all regions closed. 2024-12-09T10:50:54,027 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T10:50:54,027 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(959): stopping server 7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:54,027 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T10:50:54,027 INFO [RS:0;7ea43b39fc24:36641 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;7ea43b39fc24:36641. 2024-12-09T10:50:54,027 DEBUG [RS:0;7ea43b39fc24:36641 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:54,027 DEBUG [RS:0;7ea43b39fc24:36641 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,027 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-09T10:50:54,027 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 465444a3d0faec8193e958662a333851, disabling compactions & flushes 2024-12-09T10:50:54,027 DEBUG [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1325): Online Regions={465444a3d0faec8193e958662a333851=TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851.} 2024-12-09T10:50:54,027 INFO [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:54,028 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:54,028 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. after waiting 0 ms 2024-12-09T10:50:54,028 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:54,028 DEBUG [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1351): Waiting on 465444a3d0faec8193e958662a333851 2024-12-09T10:50:54,028 INFO [RS:1;7ea43b39fc24:46667 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T10:50:54,028 INFO [RS:1;7ea43b39fc24:46667 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T10:50:54,028 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(959): stopping server 7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:54,028 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T10:50:54,028 INFO [RS:1;7ea43b39fc24:46667 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;7ea43b39fc24:46667. 2024-12-09T10:50:54,028 DEBUG [RS:1;7ea43b39fc24:46667 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:54,028 DEBUG [RS:1;7ea43b39fc24:46667 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,029 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T10:50:54,029 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T10:50:54,029 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T10:50:54,029 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-09T10:50:54,029 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-09T10:50:54,029 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-09T10:50:54,029 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-09T10:50:54,029 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-09T10:50:54,029 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-09T10:50:54,030 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-09T10:50:54,030 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-09T10:50:54,030 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-09T10:50:54,030 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-09T10:50:54,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_1073741826_1016 (size=93) 2024-12-09T10:50:54,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_1073741826_1016 (size=93) 2024-12-09T10:50:54,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_1073741826_1016 (size=93) 2024-12-09T10:50:54,043 DEBUG [RS:2;7ea43b39fc24:33477 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/oldWALs 2024-12-09T10:50:54,043 INFO [RS:2;7ea43b39fc24:33477 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 7ea43b39fc24%2C33477%2C1733741450447:(num 1733741451763) 2024-12-09T10:50:54,043 DEBUG [RS:2;7ea43b39fc24:33477 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,043 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:54,043 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T10:50:54,043 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.ChoreService(370): Chore service for: regionserver/7ea43b39fc24:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-09T10:50:54,044 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T10:50:54,044 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T10:50:54,044 INFO [regionserver/7ea43b39fc24:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T10:50:54,044 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T10:50:54,044 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T10:50:54,044 INFO [RS:2;7ea43b39fc24:33477 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:33477 2024-12-09T10:50:54,049 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:54,049 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/7ea43b39fc24,33477,1733741450447 2024-12-09T10:50:54,050 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T10:50:54,051 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [7ea43b39fc24,33477,1733741450447] 2024-12-09T10:50:54,052 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/default/TestHBaseWalOnEC/465444a3d0faec8193e958662a333851/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-09T10:50:54,054 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/7ea43b39fc24,33477,1733741450447 already deleted, retry=false 2024-12-09T10:50:54,054 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 7ea43b39fc24,33477,1733741450447 expired; onlineServers=2 2024-12-09T10:50:54,055 INFO [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:54,055 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 465444a3d0faec8193e958662a333851: Waiting for close lock at 1733741454027Running coprocessor pre-close hooks at 1733741454027Disabling compacts and flushes for region at 1733741454027Disabling writes for close at 1733741454028 (+1 ms)Writing region close event to WAL at 1733741454029 (+1 ms)Running coprocessor post-close hooks at 1733741454053 (+24 ms)Closed at 1733741454055 (+2 ms) 2024-12-09T10:50:54,056 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851. 2024-12-09T10:50:54,064 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/info/173e552d56fa463cb6f67763895e9db1 is 153, key is TestHBaseWalOnEC,,1733741452675.465444a3d0faec8193e958662a333851./info:regioninfo/1733741453113/Put/seqid=0 2024-12-09T10:50:54,068 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,068 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,090 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-235909433_22 at /127.0.0.1:54690 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:38713:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54690 dst: /127.0.0.1:38713 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:54,096 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-09T10:50:54,097 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:54,097 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/info/173e552d56fa463cb6f67763895e9db1 2024-12-09T10:50:54,113 INFO [regionserver/7ea43b39fc24:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:54,115 INFO [regionserver/7ea43b39fc24:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:54,115 INFO [regionserver/7ea43b39fc24:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:54,140 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/ns/e178919bfa17486f8f3d65ae1a2760ea is 43, key is default/ns:d/1733741452413/Put/seqid=0 2024-12-09T10:50:54,142 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,143 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,146 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-235909433_22 at /127.0.0.1:54722 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:38713:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54722 dst: /127.0.0.1:38713 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:54,153 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-09T10:50:54,154 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:54,154 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/ns/e178919bfa17486f8f3d65ae1a2760ea 2024-12-09T10:50:54,154 INFO [RS:2;7ea43b39fc24:33477 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T10:50:54,155 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:54,155 INFO [RS:2;7ea43b39fc24:33477 {}] regionserver.HRegionServer(1031): Exiting; stopping=7ea43b39fc24,33477,1733741450447; zookeeper connection closed. 2024-12-09T10:50:54,155 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33477-0x100bd5f2d610003, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:54,155 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@6785b743 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@6785b743 2024-12-09T10:50:54,184 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/table/e3ec8e2198544be0951a8ae34ee9ef54 is 52, key is TestHBaseWalOnEC/table:state/1733741453129/Put/seqid=0 2024-12-09T10:50:54,186 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,187 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,192 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-235909433_22 at /127.0.0.1:50114 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:37003:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50114 dst: /127.0.0.1:37003 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:54,207 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-09T10:50:54,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775741_1008 (size=1189) 2024-12-09T10:50:54,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775757_1006 (size=196) 2024-12-09T10:50:54,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775756_1006 (size=196) 2024-12-09T10:50:54,211 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_-9223372036854775740_1008 (size=1189) 2024-12-09T10:50:54,214 WARN [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:54,215 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/table/e3ec8e2198544be0951a8ae34ee9ef54 2024-12-09T10:50:54,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775692_1015 (size=32) 2024-12-09T10:50:54,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775693_1015 (size=32) 2024-12-09T10:50:54,227 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/info/173e552d56fa463cb6f67763895e9db1 as hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/info/173e552d56fa463cb6f67763895e9db1 2024-12-09T10:50:54,228 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(976): stopping server 7ea43b39fc24,36641,1733741450293; all regions closed. 2024-12-09T10:50:54,229 DEBUG [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-09T10:50:54,232 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_1073741828_1018 (size=1298) 2024-12-09T10:50:54,232 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_1073741828_1018 (size=1298) 2024-12-09T10:50:54,233 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_1073741828_1018 (size=1298) 2024-12-09T10:50:54,237 DEBUG [RS:0;7ea43b39fc24:36641 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/oldWALs 2024-12-09T10:50:54,237 INFO [RS:0;7ea43b39fc24:36641 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 7ea43b39fc24%2C36641%2C1733741450293:(num 1733741451763) 2024-12-09T10:50:54,237 DEBUG [RS:0;7ea43b39fc24:36641 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,237 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:54,237 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T10:50:54,237 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.ChoreService(370): Chore service for: regionserver/7ea43b39fc24:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-09T10:50:54,238 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T10:50:54,238 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T10:50:54,238 INFO [regionserver/7ea43b39fc24:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T10:50:54,238 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T10:50:54,238 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T10:50:54,238 INFO [RS:0;7ea43b39fc24:36641 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:36641 2024-12-09T10:50:54,240 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/info/173e552d56fa463cb6f67763895e9db1, entries=10, sequenceid=11, filesize=6.5 K 2024-12-09T10:50:54,241 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/7ea43b39fc24,36641,1733741450293 2024-12-09T10:50:54,241 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T10:50:54,241 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:54,243 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [7ea43b39fc24,36641,1733741450293] 2024-12-09T10:50:54,243 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/ns/e178919bfa17486f8f3d65ae1a2760ea as hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/ns/e178919bfa17486f8f3d65ae1a2760ea 2024-12-09T10:50:54,244 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/7ea43b39fc24,36641,1733741450293 already deleted, retry=false 2024-12-09T10:50:54,244 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 7ea43b39fc24,36641,1733741450293 expired; onlineServers=1 2024-12-09T10:50:54,255 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/ns/e178919bfa17486f8f3d65ae1a2760ea, entries=2, sequenceid=11, filesize=5.0 K 2024-12-09T10:50:54,257 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/.tmp/table/e3ec8e2198544be0951a8ae34ee9ef54 as hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/table/e3ec8e2198544be0951a8ae34ee9ef54 2024-12-09T10:50:54,269 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/table/e3ec8e2198544be0951a8ae34ee9ef54, entries=2, sequenceid=11, filesize=5.1 K 2024-12-09T10:50:54,272 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 242ms, sequenceid=11, compaction requested=false 2024-12-09T10:50:54,272 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-09T10:50:54,285 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-09T10:50:54,286 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-09T10:50:54,286 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-09T10:50:54,287 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733741454029Running coprocessor pre-close hooks at 1733741454029Disabling compacts and flushes for region at 1733741454029Disabling writes for close at 1733741454030 (+1 ms)Obtaining lock to block concurrent updates at 1733741454030Preparing flush snapshotting stores in 1588230740 at 1733741454030Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733741454031 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733741454032 (+1 ms)Flushing 1588230740/info: creating writer at 1733741454032Flushing 1588230740/info: appending metadata at 1733741454060 (+28 ms)Flushing 1588230740/info: closing flushed file at 1733741454060Flushing 1588230740/ns: creating writer at 1733741454114 (+54 ms)Flushing 1588230740/ns: appending metadata at 1733741454138 (+24 ms)Flushing 1588230740/ns: closing flushed file at 1733741454138Flushing 1588230740/table: creating writer at 1733741454165 (+27 ms)Flushing 1588230740/table: appending metadata at 1733741454182 (+17 ms)Flushing 1588230740/table: closing flushed file at 1733741454182Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@79bb0db7: reopening flushed file at 1733741454225 (+43 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2b6f7047: reopening flushed file at 1733741454241 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4e819481: reopening flushed file at 1733741454255 (+14 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 242ms, sequenceid=11, compaction requested=false at 1733741454272 (+17 ms)Writing region close event to WAL at 1733741454274 (+2 ms)Running coprocessor post-close hooks at 1733741454286 (+12 ms)Closed at 1733741454286 2024-12-09T10:50:54,287 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-09T10:50:54,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775724_1010 (size=34) 2024-12-09T10:50:54,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_-9223372036854775725_1010 (size=34) 2024-12-09T10:50:54,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775772_1004 (size=42) 2024-12-09T10:50:54,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_-9223372036854775773_1004 (size=42) 2024-12-09T10:50:54,343 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:54,343 INFO [RS:0;7ea43b39fc24:36641 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T10:50:54,343 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36641-0x100bd5f2d610001, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:54,343 INFO [RS:0;7ea43b39fc24:36641 {}] regionserver.HRegionServer(1031): Exiting; stopping=7ea43b39fc24,36641,1733741450293; zookeeper connection closed. 2024-12-09T10:50:54,343 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@274df087 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@274df087 2024-12-09T10:50:54,430 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(976): stopping server 7ea43b39fc24,46667,1733741450402; all regions closed. 2024-12-09T10:50:54,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_1073741829_1019 (size=2751) 2024-12-09T10:50:54,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_1073741829_1019 (size=2751) 2024-12-09T10:50:54,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_1073741829_1019 (size=2751) 2024-12-09T10:50:54,437 DEBUG [RS:1;7ea43b39fc24:46667 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/oldWALs 2024-12-09T10:50:54,438 INFO [RS:1;7ea43b39fc24:46667 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 7ea43b39fc24%2C46667%2C1733741450402.meta:.meta(num 1733741452226) 2024-12-09T10:50:54,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_1073741827_1017 (size=93) 2024-12-09T10:50:54,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_1073741827_1017 (size=93) 2024-12-09T10:50:54,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_1073741827_1017 (size=93) 2024-12-09T10:50:54,444 DEBUG [RS:1;7ea43b39fc24:46667 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/oldWALs 2024-12-09T10:50:54,444 INFO [RS:1;7ea43b39fc24:46667 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 7ea43b39fc24%2C46667%2C1733741450402:(num 1733741451763) 2024-12-09T10:50:54,444 DEBUG [RS:1;7ea43b39fc24:46667 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:54,444 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:54,444 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T10:50:54,444 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.ChoreService(370): Chore service for: regionserver/7ea43b39fc24:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-09T10:50:54,444 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T10:50:54,444 INFO [regionserver/7ea43b39fc24:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T10:50:54,445 INFO [RS:1;7ea43b39fc24:46667 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:46667 2024-12-09T10:50:54,447 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/7ea43b39fc24,46667,1733741450402 2024-12-09T10:50:54,447 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:54,447 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T10:50:54,447 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [7ea43b39fc24,46667,1733741450402] 2024-12-09T10:50:54,449 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/7ea43b39fc24,46667,1733741450402 already deleted, retry=false 2024-12-09T10:50:54,450 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 7ea43b39fc24,46667,1733741450402 expired; onlineServers=0 2024-12-09T10:50:54,450 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '7ea43b39fc24,40433,1733741449612' ***** 2024-12-09T10:50:54,450 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-09T10:50:54,450 INFO [M:0;7ea43b39fc24:40433 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T10:50:54,450 INFO [M:0;7ea43b39fc24:40433 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T10:50:54,450 DEBUG [M:0;7ea43b39fc24:40433 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-09T10:50:54,450 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-09T10:50:54,450 DEBUG [M:0;7ea43b39fc24:40433 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-09T10:50:54,450 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.small.0-1733741451458 {}] cleaner.HFileCleaner(306): Exit Thread[master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.small.0-1733741451458,5,FailOnTimeoutGroup] 2024-12-09T10:50:54,450 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.large.0-1733741451456 {}] cleaner.HFileCleaner(306): Exit Thread[master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.large.0-1733741451456,5,FailOnTimeoutGroup] 2024-12-09T10:50:54,451 INFO [M:0;7ea43b39fc24:40433 {}] hbase.ChoreService(370): Chore service for: master/7ea43b39fc24:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-09T10:50:54,451 INFO [M:0;7ea43b39fc24:40433 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T10:50:54,451 DEBUG [M:0;7ea43b39fc24:40433 {}] master.HMaster(1795): Stopping service threads 2024-12-09T10:50:54,451 INFO [M:0;7ea43b39fc24:40433 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-09T10:50:54,451 INFO [M:0;7ea43b39fc24:40433 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-09T10:50:54,452 INFO [M:0;7ea43b39fc24:40433 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-09T10:50:54,452 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-09T10:50:54,452 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-09T10:50:54,452 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:54,452 DEBUG [M:0;7ea43b39fc24:40433 {}] zookeeper.ZKUtil(347): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-09T10:50:54,453 WARN [M:0;7ea43b39fc24:40433 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-09T10:50:54,453 INFO [M:0;7ea43b39fc24:40433 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/.lastflushedseqids 2024-12-09T10:50:54,462 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,462 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,465 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:54784 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:38713:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54784 dst: /127.0.0.1:38713 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:54,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-09T10:50:54,473 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:54,473 INFO [M:0;7ea43b39fc24:40433 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-09T10:50:54,474 INFO [M:0;7ea43b39fc24:40433 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-09T10:50:54,474 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-09T10:50:54,474 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:54,474 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:54,474 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-09T10:50:54,474 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:54,474 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.84 KB heapSize=34.13 KB 2024-12-09T10:50:54,493 DEBUG [M:0;7ea43b39fc24:40433 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/659e9937db6b465488c9acc12e994038 is 82, key is hbase:meta,,1/info:regioninfo/1733741452318/Put/seqid=0 2024-12-09T10:50:54,495 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,496 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,498 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:35618 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:37913:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35618 dst: /127.0.0.1:37913 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:54,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-09T10:50:54,503 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:54,504 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/659e9937db6b465488c9acc12e994038 2024-12-09T10:50:54,530 DEBUG [M:0;7ea43b39fc24:40433 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/07fa92007d1b404980e0a88d31555807 is 749, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733741453135/Put/seqid=0 2024-12-09T10:50:54,532 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,532 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,539 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:35634 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:37913:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35634 dst: /127.0.0.1:37913 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:54,545 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_-9223372036854775552_1037 (size=6440) 2024-12-09T10:50:54,546 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:54,546 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.16 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/07fa92007d1b404980e0a88d31555807 2024-12-09T10:50:54,549 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:54,549 INFO [RS:1;7ea43b39fc24:46667 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T10:50:54,549 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46667-0x100bd5f2d610002, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:54,549 INFO [RS:1;7ea43b39fc24:46667 {}] regionserver.HRegionServer(1031): Exiting; stopping=7ea43b39fc24,46667,1733741450402; zookeeper connection closed. 2024-12-09T10:50:54,549 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@429efae8 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@429efae8 2024-12-09T10:50:54,550 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-09T10:50:54,577 DEBUG [M:0;7ea43b39fc24:40433 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ce18f1abea5b4205a13c4b1e925ff790 is 69, key is 7ea43b39fc24,33477,1733741450447/rs:state/1733741451515/Put/seqid=0 2024-12-09T10:50:54,579 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,579 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-09T10:50:54,586 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-952322231_22 at /127.0.0.1:54808 [Receiving block BP-1773582303-172.17.0.3-1733741446266:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:38713:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54808 dst: /127.0.0.1:38713 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-09T10:50:54,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-09T10:50:54,591 WARN [M:0;7ea43b39fc24:40433 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-09T10:50:54,591 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ce18f1abea5b4205a13c4b1e925ff790 2024-12-09T10:50:54,603 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/659e9937db6b465488c9acc12e994038 as hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/659e9937db6b465488c9acc12e994038 2024-12-09T10:50:54,614 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/659e9937db6b465488c9acc12e994038, entries=8, sequenceid=72, filesize=5.5 K 2024-12-09T10:50:54,616 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/07fa92007d1b404980e0a88d31555807 as hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/07fa92007d1b404980e0a88d31555807 2024-12-09T10:50:54,625 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/07fa92007d1b404980e0a88d31555807, entries=8, sequenceid=72, filesize=6.3 K 2024-12-09T10:50:54,626 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ce18f1abea5b4205a13c4b1e925ff790 as hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/ce18f1abea5b4205a13c4b1e925ff790 2024-12-09T10:50:54,636 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/ce18f1abea5b4205a13c4b1e925ff790, entries=3, sequenceid=72, filesize=5.2 K 2024-12-09T10:50:54,638 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.84 KB/27483, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 163ms, sequenceid=72, compaction requested=false 2024-12-09T10:50:54,640 INFO [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:54,640 DEBUG [M:0;7ea43b39fc24:40433 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733741454474Disabling compacts and flushes for region at 1733741454474Disabling writes for close at 1733741454474Obtaining lock to block concurrent updates at 1733741454474Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733741454474Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27483, getHeapSize=34880, getOffHeapSize=0, getCellsCount=85 at 1733741454475 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733741454476 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733741454476Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733741454493 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733741454493Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733741454512 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733741454529 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733741454529Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733741454555 (+26 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733741454576 (+21 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733741454576Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3fa3fbbd: reopening flushed file at 1733741454601 (+25 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@54ea9a3b: reopening flushed file at 1733741454614 (+13 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3ac0eb0: reopening flushed file at 1733741454625 (+11 ms)Finished flush of dataSize ~26.84 KB/27483, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 163ms, sequenceid=72, compaction requested=false at 1733741454638 (+13 ms)Writing region close event to WAL at 1733741454640 (+2 ms)Closed at 1733741454640 2024-12-09T10:50:54,644 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38713 is added to blk_1073741825_1011 (size=32686) 2024-12-09T10:50:54,645 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37003 is added to blk_1073741825_1011 (size=32686) 2024-12-09T10:50:54,645 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37913 is added to blk_1073741825_1011 (size=32686) 2024-12-09T10:50:54,646 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T10:50:54,646 INFO [M:0;7ea43b39fc24:40433 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-09T10:50:54,646 INFO [M:0;7ea43b39fc24:40433 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:40433 2024-12-09T10:50:54,646 INFO [M:0;7ea43b39fc24:40433 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T10:50:54,749 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:54,749 INFO [M:0;7ea43b39fc24:40433 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T10:50:54,749 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40433-0x100bd5f2d610000, quorum=127.0.0.1:58213, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:54,753 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2e59159d{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:54,756 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@a8e922f{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T10:50:54,756 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T10:50:54,756 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@24f92c39{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T10:50:54,756 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@c62369b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir/,STOPPED} 2024-12-09T10:50:54,759 WARN [BP-1773582303-172.17.0.3-1733741446266 heartbeating to localhost/127.0.0.1:33501 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T10:50:54,759 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T10:50:54,759 WARN [BP-1773582303-172.17.0.3-1733741446266 heartbeating to localhost/127.0.0.1:33501 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1773582303-172.17.0.3-1733741446266 (Datanode Uuid b11aa68a-130a-4d14-902b-69cf740f7d6c) service to localhost/127.0.0.1:33501 2024-12-09T10:50:54,759 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T10:50:54,760 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data5/current/BP-1773582303-172.17.0.3-1733741446266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:54,760 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data6/current/BP-1773582303-172.17.0.3-1733741446266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:54,761 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T10:50:54,763 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1c6b8f01{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:54,763 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11f28dd2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T10:50:54,764 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T10:50:54,764 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7fa8fa5c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T10:50:54,764 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6463ad04{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir/,STOPPED} 2024-12-09T10:50:54,765 WARN [BP-1773582303-172.17.0.3-1733741446266 heartbeating to localhost/127.0.0.1:33501 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T10:50:54,765 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T10:50:54,765 WARN [BP-1773582303-172.17.0.3-1733741446266 heartbeating to localhost/127.0.0.1:33501 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1773582303-172.17.0.3-1733741446266 (Datanode Uuid fef88b2f-b468-4668-9d86-df610b6e1f0e) service to localhost/127.0.0.1:33501 2024-12-09T10:50:54,765 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T10:50:54,766 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data3/current/BP-1773582303-172.17.0.3-1733741446266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:54,766 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data4/current/BP-1773582303-172.17.0.3-1733741446266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:54,766 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T10:50:54,770 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4839957b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:54,770 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5306f615{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T10:50:54,771 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T10:50:54,771 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1a2478ad{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T10:50:54,771 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@550154bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir/,STOPPED} 2024-12-09T10:50:54,772 WARN [BP-1773582303-172.17.0.3-1733741446266 heartbeating to localhost/127.0.0.1:33501 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T10:50:54,772 WARN [BP-1773582303-172.17.0.3-1733741446266 heartbeating to localhost/127.0.0.1:33501 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1773582303-172.17.0.3-1733741446266 (Datanode Uuid c8633e2b-43b2-4753-8101-90f583d169cb) service to localhost/127.0.0.1:33501 2024-12-09T10:50:54,773 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data1/current/BP-1773582303-172.17.0.3-1733741446266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:54,773 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/cluster_cbca2f39-76bc-87c6-1d3e-9c33de5cdde3/data/data2/current/BP-1773582303-172.17.0.3-1733741446266 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:54,773 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T10:50:54,773 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T10:50:54,773 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T10:50:54,782 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@76e4c45c{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-09T10:50:54,783 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T10:50:54,783 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T10:50:54,783 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T10:50:54,783 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir/,STOPPED} 2024-12-09T10:50:54,793 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-09T10:50:54,822 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-09T10:50:54,831 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=92 (was 159), OpenFileDescriptor=441 (was 391) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=199 (was 198) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=7770 (was 8085) 2024-12-09T10:50:54,838 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=92, OpenFileDescriptor=441, MaxFileDescriptor=1048576, SystemLoadAverage=199, ProcessCount=11, AvailableMemoryMB=7769 2024-12-09T10:50:54,838 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-09T10:50:54,838 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.log.dir so I do NOT create it in target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4bdc1f7b-4242-49fb-400c-456d657659e0/hadoop.tmp.dir so I do NOT create it in target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1, deleteOnExit=true 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/test.cache.data in system properties and HBase conf 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.tmp.dir in system properties and HBase conf 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir in system properties and HBase conf 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-09T10:50:54,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-09T10:50:54,839 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/nfs.dump.dir in system properties and HBase conf 2024-12-09T10:50:54,840 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/java.io.tmpdir in system properties and HBase conf 2024-12-09T10:50:54,841 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-09T10:50:54,841 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-09T10:50:54,841 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-09T10:50:54,932 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:54,937 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T10:50:54,938 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T10:50:54,938 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T10:50:54,938 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-09T10:50:54,939 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:54,939 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@715f09c8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir/,AVAILABLE} 2024-12-09T10:50:54,940 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@71b7cabb{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T10:50:55,057 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@46039787{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/java.io.tmpdir/jetty-localhost-36133-hadoop-hdfs-3_4_1-tests_jar-_-any-11726336478663451587/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-09T10:50:55,058 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@7326bb42{HTTP/1.1, (http/1.1)}{localhost:36133} 2024-12-09T10:50:55,058 INFO [Time-limited test {}] server.Server(415): Started @10829ms 2024-12-09T10:50:55,141 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:55,144 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T10:50:55,145 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T10:50:55,145 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T10:50:55,145 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-09T10:50:55,147 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7c2dd4e6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir/,AVAILABLE} 2024-12-09T10:50:55,147 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2591ff9a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T10:50:55,262 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@38845bbf{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/java.io.tmpdir/jetty-localhost-32999-hadoop-hdfs-3_4_1-tests_jar-_-any-16762095420366864683/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:55,263 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@24815bd8{HTTP/1.1, (http/1.1)}{localhost:32999} 2024-12-09T10:50:55,263 INFO [Time-limited test {}] server.Server(415): Started @11034ms 2024-12-09T10:50:55,264 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T10:50:55,299 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:55,303 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T10:50:55,305 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T10:50:55,305 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T10:50:55,305 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-09T10:50:55,306 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6f8ad177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir/,AVAILABLE} 2024-12-09T10:50:55,306 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1fee469f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T10:50:55,353 WARN [Thread-526 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data1/current/BP-1261905413-172.17.0.3-1733741454874/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:55,353 WARN [Thread-527 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data2/current/BP-1261905413-172.17.0.3-1733741454874/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:55,371 WARN [Thread-505 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T10:50:55,374 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xed3598c55ea99ece with lease ID 0x32541052a606a436: Processing first storage report for DS-919eb513-ba53-4055-a086-3dd81df44291 from datanode DatanodeRegistration(127.0.0.1:36347, datanodeUuid=17ec257c-95f6-41e5-98a3-f3eed91d4d86, infoPort=40793, infoSecurePort=0, ipcPort=33267, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874) 2024-12-09T10:50:55,374 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xed3598c55ea99ece with lease ID 0x32541052a606a436: from storage DS-919eb513-ba53-4055-a086-3dd81df44291 node DatanodeRegistration(127.0.0.1:36347, datanodeUuid=17ec257c-95f6-41e5-98a3-f3eed91d4d86, infoPort=40793, infoSecurePort=0, ipcPort=33267, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T10:50:55,374 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xed3598c55ea99ece with lease ID 0x32541052a606a436: Processing first storage report for DS-0306aac6-a5c6-46f4-add1-122b97180e23 from datanode DatanodeRegistration(127.0.0.1:36347, datanodeUuid=17ec257c-95f6-41e5-98a3-f3eed91d4d86, infoPort=40793, infoSecurePort=0, ipcPort=33267, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874) 2024-12-09T10:50:55,374 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xed3598c55ea99ece with lease ID 0x32541052a606a436: from storage DS-0306aac6-a5c6-46f4-add1-122b97180e23 node DatanodeRegistration(127.0.0.1:36347, datanodeUuid=17ec257c-95f6-41e5-98a3-f3eed91d4d86, infoPort=40793, infoSecurePort=0, ipcPort=33267, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T10:50:55,426 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@18a67058{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/java.io.tmpdir/jetty-localhost-46769-hadoop-hdfs-3_4_1-tests_jar-_-any-10283713983041611128/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:55,427 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@478131f{HTTP/1.1, (http/1.1)}{localhost:46769} 2024-12-09T10:50:55,427 INFO [Time-limited test {}] server.Server(415): Started @11198ms 2024-12-09T10:50:55,428 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T10:50:55,458 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-09T10:50:55,461 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-09T10:50:55,462 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-09T10:50:55,462 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-09T10:50:55,462 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-09T10:50:55,463 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@739551bc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir/,AVAILABLE} 2024-12-09T10:50:55,463 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@19f40ccf{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-09T10:50:55,513 WARN [Thread-561 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data3/current/BP-1261905413-172.17.0.3-1733741454874/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:55,513 WARN [Thread-562 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data4/current/BP-1261905413-172.17.0.3-1733741454874/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:55,530 WARN [Thread-541 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T10:50:55,534 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe3fe0bd16439c37c with lease ID 0x32541052a606a437: Processing first storage report for DS-08f1da18-d862-41a6-8296-a2e287df4e71 from datanode DatanodeRegistration(127.0.0.1:36897, datanodeUuid=6dc06157-63cc-42b2-b5b3-4e205a2b6465, infoPort=37361, infoSecurePort=0, ipcPort=46365, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874) 2024-12-09T10:50:55,534 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe3fe0bd16439c37c with lease ID 0x32541052a606a437: from storage DS-08f1da18-d862-41a6-8296-a2e287df4e71 node DatanodeRegistration(127.0.0.1:36897, datanodeUuid=6dc06157-63cc-42b2-b5b3-4e205a2b6465, infoPort=37361, infoSecurePort=0, ipcPort=46365, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-09T10:50:55,534 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe3fe0bd16439c37c with lease ID 0x32541052a606a437: Processing first storage report for DS-f6460373-a9da-46b4-8011-9822b3a8fbdd from datanode DatanodeRegistration(127.0.0.1:36897, datanodeUuid=6dc06157-63cc-42b2-b5b3-4e205a2b6465, infoPort=37361, infoSecurePort=0, ipcPort=46365, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874) 2024-12-09T10:50:55,534 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe3fe0bd16439c37c with lease ID 0x32541052a606a437: from storage DS-f6460373-a9da-46b4-8011-9822b3a8fbdd node DatanodeRegistration(127.0.0.1:36897, datanodeUuid=6dc06157-63cc-42b2-b5b3-4e205a2b6465, infoPort=37361, infoSecurePort=0, ipcPort=46365, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T10:50:55,581 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@472e7194{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/java.io.tmpdir/jetty-localhost-33969-hadoop-hdfs-3_4_1-tests_jar-_-any-1133268336946441830/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:55,581 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@7ba9a3d5{HTTP/1.1, (http/1.1)}{localhost:33969} 2024-12-09T10:50:55,581 INFO [Time-limited test {}] server.Server(415): Started @11352ms 2024-12-09T10:50:55,583 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-09T10:50:55,668 WARN [Thread-587 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data5/current/BP-1261905413-172.17.0.3-1733741454874/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:55,668 WARN [Thread-588 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data6/current/BP-1261905413-172.17.0.3-1733741454874/current, will proceed with Du for space computation calculation, 2024-12-09T10:50:55,685 WARN [Thread-576 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-09T10:50:55,688 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xf5a04711dbec5c0f with lease ID 0x32541052a606a438: Processing first storage report for DS-9694783b-4da0-4299-8133-36270ba395e9 from datanode DatanodeRegistration(127.0.0.1:44245, datanodeUuid=f7633a3e-9018-4036-a2e6-39607bb150ce, infoPort=43387, infoSecurePort=0, ipcPort=44365, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874) 2024-12-09T10:50:55,688 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xf5a04711dbec5c0f with lease ID 0x32541052a606a438: from storage DS-9694783b-4da0-4299-8133-36270ba395e9 node DatanodeRegistration(127.0.0.1:44245, datanodeUuid=f7633a3e-9018-4036-a2e6-39607bb150ce, infoPort=43387, infoSecurePort=0, ipcPort=44365, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-09T10:50:55,689 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xf5a04711dbec5c0f with lease ID 0x32541052a606a438: Processing first storage report for DS-a37f6df9-67df-4427-99f7-75b7bc12c1bc from datanode DatanodeRegistration(127.0.0.1:44245, datanodeUuid=f7633a3e-9018-4036-a2e6-39607bb150ce, infoPort=43387, infoSecurePort=0, ipcPort=44365, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874) 2024-12-09T10:50:55,689 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xf5a04711dbec5c0f with lease ID 0x32541052a606a438: from storage DS-a37f6df9-67df-4427-99f7-75b7bc12c1bc node DatanodeRegistration(127.0.0.1:44245, datanodeUuid=f7633a3e-9018-4036-a2e6-39607bb150ce, infoPort=43387, infoSecurePort=0, ipcPort=44365, storageInfo=lv=-57;cid=testClusterID;nsid=571447898;c=1733741454874), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-09T10:50:55,709 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b 2024-12-09T10:50:55,712 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/zookeeper_0, clientPort=59671, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-09T10:50:55,713 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=59671 2024-12-09T10:50:55,713 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,715 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741825_1001 (size=7) 2024-12-09T10:50:55,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741825_1001 (size=7) 2024-12-09T10:50:55,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741825_1001 (size=7) 2024-12-09T10:50:55,730 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def with version=8 2024-12-09T10:50:55,731 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:33501/user/jenkins/test-data/cfd3e2dc-2560-3ed9-bb05-4682b964c8db/hbase-staging 2024-12-09T10:50:55,733 INFO [Time-limited test {}] client.ConnectionUtils(128): master/7ea43b39fc24:0 server-side Connection retries=45 2024-12-09T10:50:55,733 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,733 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,733 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T10:50:55,733 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,733 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T10:50:55,733 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-09T10:50:55,733 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T10:50:55,734 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:46161 2024-12-09T10:50:55,735 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:46161 connecting to ZooKeeper ensemble=127.0.0.1:59671 2024-12-09T10:50:55,741 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:461610x0, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T10:50:55,742 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:46161-0x100bd5f48660000 connected 2024-12-09T10:50:55,756 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,758 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,760 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:55,760 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def, hbase.cluster.distributed=false 2024-12-09T10:50:55,762 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T10:50:55,762 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46161 2024-12-09T10:50:55,762 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46161 2024-12-09T10:50:55,763 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46161 2024-12-09T10:50:55,763 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46161 2024-12-09T10:50:55,764 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46161 2024-12-09T10:50:55,779 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/7ea43b39fc24:0 server-side Connection retries=45 2024-12-09T10:50:55,779 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,779 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,779 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T10:50:55,779 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,779 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T10:50:55,779 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T10:50:55,780 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T10:50:55,780 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:40291 2024-12-09T10:50:55,781 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:40291 connecting to ZooKeeper ensemble=127.0.0.1:59671 2024-12-09T10:50:55,782 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,785 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,789 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:402910x0, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T10:50:55,789 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:402910x0, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:55,790 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:40291-0x100bd5f48660001 connected 2024-12-09T10:50:55,790 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T10:50:55,790 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T10:50:55,791 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T10:50:55,792 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T10:50:55,793 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40291 2024-12-09T10:50:55,793 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40291 2024-12-09T10:50:55,793 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40291 2024-12-09T10:50:55,794 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40291 2024-12-09T10:50:55,794 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40291 2024-12-09T10:50:55,810 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/7ea43b39fc24:0 server-side Connection retries=45 2024-12-09T10:50:55,810 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,810 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,810 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T10:50:55,810 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,810 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T10:50:55,810 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T10:50:55,810 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T10:50:55,811 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:39389 2024-12-09T10:50:55,812 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39389 connecting to ZooKeeper ensemble=127.0.0.1:59671 2024-12-09T10:50:55,813 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,815 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:393890x0, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T10:50:55,820 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39389-0x100bd5f48660002 connected 2024-12-09T10:50:55,820 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:55,820 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T10:50:55,821 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T10:50:55,821 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T10:50:55,823 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T10:50:55,823 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39389 2024-12-09T10:50:55,823 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39389 2024-12-09T10:50:55,823 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39389 2024-12-09T10:50:55,824 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39389 2024-12-09T10:50:55,824 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39389 2024-12-09T10:50:55,839 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/7ea43b39fc24:0 server-side Connection retries=45 2024-12-09T10:50:55,839 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,839 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,840 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-09T10:50:55,840 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-09T10:50:55,840 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-09T10:50:55,840 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-09T10:50:55,840 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-09T10:50:55,840 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:45835 2024-12-09T10:50:55,842 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:45835 connecting to ZooKeeper ensemble=127.0.0.1:59671 2024-12-09T10:50:55,842 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,844 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,848 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:458350x0, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-09T10:50:55,849 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:458350x0, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:55,849 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45835-0x100bd5f48660003 connected 2024-12-09T10:50:55,849 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-09T10:50:55,850 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-09T10:50:55,850 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-09T10:50:55,851 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-09T10:50:55,852 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45835 2024-12-09T10:50:55,852 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45835 2024-12-09T10:50:55,852 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45835 2024-12-09T10:50:55,853 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45835 2024-12-09T10:50:55,853 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45835 2024-12-09T10:50:55,866 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;7ea43b39fc24:46161 2024-12-09T10:50:55,866 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:55,869 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:55,869 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:55,869 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:55,869 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:55,872 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:55,874 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T10:50:55,874 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:55,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:55,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T10:50:55,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-09T10:50:55,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:55,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:55,875 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-09T10:50:55,876 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/7ea43b39fc24,46161,1733741455732 from backup master directory 2024-12-09T10:50:55,879 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:55,879 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:55,879 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:55,879 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:55,879 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T10:50:55,879 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-09T10:50:55,879 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:55,885 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/hbase.id] with ID: a99fdc1b-408b-4818-9d28-02d317339801 2024-12-09T10:50:55,885 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/.tmp/hbase.id 2024-12-09T10:50:55,894 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741826_1002 (size=42) 2024-12-09T10:50:55,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741826_1002 (size=42) 2024-12-09T10:50:55,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741826_1002 (size=42) 2024-12-09T10:50:55,896 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/.tmp/hbase.id]:[hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/hbase.id] 2024-12-09T10:50:55,912 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-09T10:50:55,912 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-09T10:50:55,914 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-12-09T10:50:55,916 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:55,916 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:55,916 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:55,916 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:55,926 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741827_1003 (size=196) 2024-12-09T10:50:55,926 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741827_1003 (size=196) 2024-12-09T10:50:55,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741827_1003 (size=196) 2024-12-09T10:50:55,927 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-09T10:50:55,928 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-09T10:50:55,928 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T10:50:55,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741828_1004 (size=1189) 2024-12-09T10:50:55,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741828_1004 (size=1189) 2024-12-09T10:50:55,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741828_1004 (size=1189) 2024-12-09T10:50:55,942 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store 2024-12-09T10:50:55,952 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741829_1005 (size=34) 2024-12-09T10:50:55,952 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741829_1005 (size=34) 2024-12-09T10:50:55,953 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741829_1005 (size=34) 2024-12-09T10:50:55,953 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:55,953 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-09T10:50:55,953 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:55,954 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:55,954 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-09T10:50:55,954 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:55,954 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:55,954 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733741455953Disabling compacts and flushes for region at 1733741455953Disabling writes for close at 1733741455954 (+1 ms)Writing region close event to WAL at 1733741455954Closed at 1733741455954 2024-12-09T10:50:55,955 WARN [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/.initializing 2024-12-09T10:50:55,955 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/WALs/7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:55,961 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C46161%2C1733741455732, suffix=, logDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/WALs/7ea43b39fc24,46161,1733741455732, archiveDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/oldWALs, maxLogs=10 2024-12-09T10:50:55,962 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 7ea43b39fc24%2C46161%2C1733741455732.1733741455961 2024-12-09T10:50:55,973 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/WALs/7ea43b39fc24,46161,1733741455732/7ea43b39fc24%2C46161%2C1733741455732.1733741455961 2024-12-09T10:50:55,976 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40793:40793),(127.0.0.1/127.0.0.1:43387:43387),(127.0.0.1/127.0.0.1:37361:37361)] 2024-12-09T10:50:55,979 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-09T10:50:55,979 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:55,979 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,979 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,982 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,983 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-09T10:50:55,984 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:55,984 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:55,984 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,986 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-09T10:50:55,986 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:55,987 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T10:50:55,987 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,989 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-09T10:50:55,990 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:55,990 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T10:50:55,990 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,992 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-09T10:50:55,992 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:55,993 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T10:50:55,993 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,994 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,995 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,996 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,996 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:55,997 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T10:50:55,999 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-09T10:50:56,001 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T10:50:56,002 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61878345, jitterRate=-0.07794080674648285}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T10:50:56,003 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733741455979Initializing all the Stores at 1733741455981 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741455981Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741455981Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741455981Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741455982 (+1 ms)Cleaning up temporary data from old regions at 1733741455996 (+14 ms)Region opened successfully at 1733741456003 (+7 ms) 2024-12-09T10:50:56,005 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-09T10:50:56,009 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7197aae8, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=7ea43b39fc24/172.17.0.3:0 2024-12-09T10:50:56,010 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-09T10:50:56,011 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-09T10:50:56,011 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-09T10:50:56,011 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-09T10:50:56,011 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-09T10:50:56,012 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-09T10:50:56,012 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-09T10:50:56,014 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-09T10:50:56,015 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-09T10:50:56,017 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-09T10:50:56,017 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-09T10:50:56,018 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-09T10:50:56,019 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-09T10:50:56,019 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-09T10:50:56,020 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-09T10:50:56,022 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-09T10:50:56,023 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-09T10:50:56,025 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-09T10:50:56,027 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-09T10:50:56,029 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-09T10:50:56,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:56,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:56,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:56,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:56,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,032 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=7ea43b39fc24,46161,1733741455732, sessionid=0x100bd5f48660000, setting cluster-up flag (Was=false) 2024-12-09T10:50:56,035 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,035 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,035 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,035 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,041 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-09T10:50:56,042 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:56,046 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,046 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,046 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,046 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,051 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-09T10:50:56,052 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:56,054 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-09T10:50:56,057 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-09T10:50:56,057 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-09T10:50:56,057 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-09T10:50:56,058 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 7ea43b39fc24,46161,1733741455732 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-09T10:50:56,059 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/7ea43b39fc24:0, corePoolSize=5, maxPoolSize=5 2024-12-09T10:50:56,059 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/7ea43b39fc24:0, corePoolSize=5, maxPoolSize=5 2024-12-09T10:50:56,059 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=5, maxPoolSize=5 2024-12-09T10:50:56,059 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=5, maxPoolSize=5 2024-12-09T10:50:56,059 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/7ea43b39fc24:0, corePoolSize=10, maxPoolSize=10 2024-12-09T10:50:56,060 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,060 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=2, maxPoolSize=2 2024-12-09T10:50:56,060 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,062 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T10:50:56,062 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-09T10:50:56,064 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,064 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-09T10:50:56,066 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733741486065 2024-12-09T10:50:56,066 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-09T10:50:56,066 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-09T10:50:56,066 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-09T10:50:56,066 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-09T10:50:56,066 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-09T10:50:56,066 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-09T10:50:56,066 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,067 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-09T10:50:56,067 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-09T10:50:56,067 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-09T10:50:56,069 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-09T10:50:56,069 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-09T10:50:56,069 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.large.0-1733741456069,5,FailOnTimeoutGroup] 2024-12-09T10:50:56,069 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.small.0-1733741456069,5,FailOnTimeoutGroup] 2024-12-09T10:50:56,070 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,070 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-09T10:50:56,070 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,070 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,078 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741831_1007 (size=1321) 2024-12-09T10:50:56,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741831_1007 (size=1321) 2024-12-09T10:50:56,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741831_1007 (size=1321) 2024-12-09T10:50:56,080 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-09T10:50:56,080 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def 2024-12-09T10:50:56,092 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741832_1008 (size=32) 2024-12-09T10:50:56,092 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741832_1008 (size=32) 2024-12-09T10:50:56,093 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741832_1008 (size=32) 2024-12-09T10:50:56,094 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:56,095 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-09T10:50:56,097 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-09T10:50:56,097 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,098 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:56,098 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-09T10:50:56,099 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-09T10:50:56,099 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,100 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:56,100 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-09T10:50:56,101 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-09T10:50:56,102 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,102 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:56,102 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-09T10:50:56,104 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-09T10:50:56,104 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,104 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:56,105 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-09T10:50:56,105 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740 2024-12-09T10:50:56,106 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740 2024-12-09T10:50:56,108 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-09T10:50:56,108 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-09T10:50:56,108 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T10:50:56,110 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-09T10:50:56,112 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T10:50:56,113 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64256402, jitterRate=-0.04250499606132507}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T10:50:56,114 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733741456094Initializing all the Stores at 1733741456095 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741456095Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741456095Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741456095Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741456095Cleaning up temporary data from old regions at 1733741456108 (+13 ms)Region opened successfully at 1733741456114 (+6 ms) 2024-12-09T10:50:56,114 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-09T10:50:56,114 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-09T10:50:56,114 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-09T10:50:56,114 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-09T10:50:56,114 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-09T10:50:56,114 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-09T10:50:56,114 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733741456114Disabling compacts and flushes for region at 1733741456114Disabling writes for close at 1733741456114Writing region close event to WAL at 1733741456114Closed at 1733741456114 2024-12-09T10:50:56,116 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T10:50:56,117 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-09T10:50:56,117 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-09T10:50:56,119 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-09T10:50:56,121 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-09T10:50:56,155 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(746): ClusterId : a99fdc1b-408b-4818-9d28-02d317339801 2024-12-09T10:50:56,155 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(746): ClusterId : a99fdc1b-408b-4818-9d28-02d317339801 2024-12-09T10:50:56,155 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(746): ClusterId : a99fdc1b-408b-4818-9d28-02d317339801 2024-12-09T10:50:56,155 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T10:50:56,155 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T10:50:56,155 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-09T10:50:56,158 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T10:50:56,158 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T10:50:56,158 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-09T10:50:56,158 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T10:50:56,158 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T10:50:56,158 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-09T10:50:56,161 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T10:50:56,161 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T10:50:56,161 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-09T10:50:56,161 DEBUG [RS:1;7ea43b39fc24:39389 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2f35148e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=7ea43b39fc24/172.17.0.3:0 2024-12-09T10:50:56,161 DEBUG [RS:0;7ea43b39fc24:40291 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@bd56829, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=7ea43b39fc24/172.17.0.3:0 2024-12-09T10:50:56,161 DEBUG [RS:2;7ea43b39fc24:45835 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7ebec686, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=7ea43b39fc24/172.17.0.3:0 2024-12-09T10:50:56,175 DEBUG [RS:1;7ea43b39fc24:39389 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;7ea43b39fc24:39389 2024-12-09T10:50:56,175 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T10:50:56,175 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T10:50:56,175 DEBUG [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T10:50:56,176 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(2659): reportForDuty to master=7ea43b39fc24,46161,1733741455732 with port=39389, startcode=1733741455809 2024-12-09T10:50:56,177 DEBUG [RS:1;7ea43b39fc24:39389 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T10:50:56,179 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37279, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T10:50:56,179 DEBUG [RS:0;7ea43b39fc24:40291 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;7ea43b39fc24:40291 2024-12-09T10:50:56,180 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T10:50:56,180 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T10:50:56,180 DEBUG [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T10:50:56,180 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46161 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 7ea43b39fc24,39389,1733741455809 2024-12-09T10:50:56,180 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46161 {}] master.ServerManager(517): Registering regionserver=7ea43b39fc24,39389,1733741455809 2024-12-09T10:50:56,180 DEBUG [RS:2;7ea43b39fc24:45835 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;7ea43b39fc24:45835 2024-12-09T10:50:56,180 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-09T10:50:56,181 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-09T10:50:56,181 DEBUG [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-09T10:50:56,181 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(2659): reportForDuty to master=7ea43b39fc24,46161,1733741455732 with port=40291, startcode=1733741455779 2024-12-09T10:50:56,181 DEBUG [RS:0;7ea43b39fc24:40291 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T10:50:56,181 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(2659): reportForDuty to master=7ea43b39fc24,46161,1733741455732 with port=45835, startcode=1733741455839 2024-12-09T10:50:56,182 DEBUG [RS:2;7ea43b39fc24:45835 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-09T10:50:56,183 DEBUG [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def 2024-12-09T10:50:56,183 DEBUG [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:34523 2024-12-09T10:50:56,183 DEBUG [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T10:50:56,184 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:58473, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T10:50:56,184 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37011, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-09T10:50:56,185 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46161 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:56,185 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46161 {}] master.ServerManager(517): Registering regionserver=7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:56,185 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:56,186 DEBUG [RS:1;7ea43b39fc24:39389 {}] zookeeper.ZKUtil(111): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/7ea43b39fc24,39389,1733741455809 2024-12-09T10:50:56,186 WARN [RS:1;7ea43b39fc24:39389 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T10:50:56,186 INFO [RS:1;7ea43b39fc24:39389 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T10:50:56,186 DEBUG [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,39389,1733741455809 2024-12-09T10:50:56,187 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [7ea43b39fc24,39389,1733741455809] 2024-12-09T10:50:56,187 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46161 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,188 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46161 {}] master.ServerManager(517): Registering regionserver=7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,188 DEBUG [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def 2024-12-09T10:50:56,188 DEBUG [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:34523 2024-12-09T10:50:56,188 DEBUG [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T10:50:56,190 DEBUG [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def 2024-12-09T10:50:56,190 DEBUG [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:34523 2024-12-09T10:50:56,190 DEBUG [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-09T10:50:56,192 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:56,194 DEBUG [RS:2;7ea43b39fc24:45835 {}] zookeeper.ZKUtil(111): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:56,194 WARN [RS:2;7ea43b39fc24:45835 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T10:50:56,194 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [7ea43b39fc24,45835,1733741455839] 2024-12-09T10:50:56,194 INFO [RS:2;7ea43b39fc24:45835 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T10:50:56,194 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [7ea43b39fc24,40291,1733741455779] 2024-12-09T10:50:56,194 DEBUG [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:56,194 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T10:50:56,195 DEBUG [RS:0;7ea43b39fc24:40291 {}] zookeeper.ZKUtil(111): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,195 WARN [RS:0;7ea43b39fc24:40291 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-09T10:50:56,195 INFO [RS:0;7ea43b39fc24:40291 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T10:50:56,195 DEBUG [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,197 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T10:50:56,197 INFO [RS:1;7ea43b39fc24:39389 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T10:50:56,197 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,198 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T10:50:56,199 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T10:50:56,199 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,199 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T10:50:56,200 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/7ea43b39fc24:0, corePoolSize=2, maxPoolSize=2 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,200 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,201 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,201 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:56,201 DEBUG [RS:1;7ea43b39fc24:39389 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:56,201 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,201 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,202 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,202 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,202 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T10:50:56,202 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,202 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,39389,1733741455809-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T10:50:56,206 INFO [RS:2;7ea43b39fc24:45835 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T10:50:56,206 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,207 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-09T10:50:56,208 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T10:50:56,208 INFO [RS:0;7ea43b39fc24:40291 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-09T10:50:56,208 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,209 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-09T10:50:56,209 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T10:50:56,209 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,209 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,209 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,209 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,209 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,209 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-09T10:50:56,209 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,209 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/7ea43b39fc24:0, corePoolSize=2, maxPoolSize=2 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/7ea43b39fc24:0, corePoolSize=2, maxPoolSize=2 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/7ea43b39fc24:0, corePoolSize=1, maxPoolSize=1 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:56,210 DEBUG [RS:2;7ea43b39fc24:45835 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:56,210 DEBUG [RS:0;7ea43b39fc24:40291 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0, corePoolSize=3, maxPoolSize=3 2024-12-09T10:50:56,213 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,45835,1733741455839-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T10:50:56,213 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,40291,1733741455779-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T10:50:56,230 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T10:50:56,230 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T10:50:56,230 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,39389,1733741455809-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,231 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,40291,1733741455779-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,231 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,231 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,231 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.Replication(171): 7ea43b39fc24,39389,1733741455809 started 2024-12-09T10:50:56,231 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.Replication(171): 7ea43b39fc24,40291,1733741455779 started 2024-12-09T10:50:56,232 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-09T10:50:56,232 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,45835,1733741455839-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,232 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,232 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.Replication(171): 7ea43b39fc24,45835,1733741455839 started 2024-12-09T10:50:56,248 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,248 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,248 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1482): Serving as 7ea43b39fc24,45835,1733741455839, RpcServer on 7ea43b39fc24/172.17.0.3:45835, sessionid=0x100bd5f48660003 2024-12-09T10:50:56,248 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1482): Serving as 7ea43b39fc24,40291,1733741455779, RpcServer on 7ea43b39fc24/172.17.0.3:40291, sessionid=0x100bd5f48660001 2024-12-09T10:50:56,248 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T10:50:56,248 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T10:50:56,248 DEBUG [RS:0;7ea43b39fc24:40291 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,248 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,40291,1733741455779' 2024-12-09T10:50:56,248 DEBUG [RS:2;7ea43b39fc24:45835 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:56,248 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T10:50:56,248 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,45835,1733741455839' 2024-12-09T10:50:56,248 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T10:50:56,249 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T10:50:56,249 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T10:50:56,250 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T10:50:56,250 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T10:50:56,250 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T10:50:56,250 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T10:50:56,250 DEBUG [RS:0;7ea43b39fc24:40291 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,250 DEBUG [RS:2;7ea43b39fc24:45835 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:56,250 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,40291,1733741455779' 2024-12-09T10:50:56,250 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,45835,1733741455839' 2024-12-09T10:50:56,250 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T10:50:56,250 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T10:50:56,250 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T10:50:56,250 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T10:50:56,251 DEBUG [RS:0;7ea43b39fc24:40291 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T10:50:56,251 DEBUG [RS:2;7ea43b39fc24:45835 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T10:50:56,251 INFO [RS:0;7ea43b39fc24:40291 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T10:50:56,251 INFO [RS:2;7ea43b39fc24:45835 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T10:50:56,251 INFO [RS:0;7ea43b39fc24:40291 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T10:50:56,251 INFO [RS:2;7ea43b39fc24:45835 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T10:50:56,252 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,253 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(1482): Serving as 7ea43b39fc24,39389,1733741455809, RpcServer on 7ea43b39fc24/172.17.0.3:39389, sessionid=0x100bd5f48660002 2024-12-09T10:50:56,253 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-09T10:50:56,253 DEBUG [RS:1;7ea43b39fc24:39389 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 7ea43b39fc24,39389,1733741455809 2024-12-09T10:50:56,253 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,39389,1733741455809' 2024-12-09T10:50:56,253 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-09T10:50:56,253 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-09T10:50:56,254 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-09T10:50:56,254 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-09T10:50:56,254 DEBUG [RS:1;7ea43b39fc24:39389 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 7ea43b39fc24,39389,1733741455809 2024-12-09T10:50:56,254 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '7ea43b39fc24,39389,1733741455809' 2024-12-09T10:50:56,254 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-09T10:50:56,255 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-09T10:50:56,255 DEBUG [RS:1;7ea43b39fc24:39389 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-09T10:50:56,255 INFO [RS:1;7ea43b39fc24:39389 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-09T10:50:56,255 INFO [RS:1;7ea43b39fc24:39389 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-09T10:50:56,271 WARN [7ea43b39fc24:46161 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-09T10:50:56,354 INFO [RS:0;7ea43b39fc24:40291 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C40291%2C1733741455779, suffix=, logDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,40291,1733741455779, archiveDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/oldWALs, maxLogs=32 2024-12-09T10:50:56,354 INFO [RS:2;7ea43b39fc24:45835 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C45835%2C1733741455839, suffix=, logDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,45835,1733741455839, archiveDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/oldWALs, maxLogs=32 2024-12-09T10:50:56,356 INFO [RS:2;7ea43b39fc24:45835 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 7ea43b39fc24%2C45835%2C1733741455839.1733741456356 2024-12-09T10:50:56,356 INFO [RS:0;7ea43b39fc24:40291 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 7ea43b39fc24%2C40291%2C1733741455779.1733741456356 2024-12-09T10:50:56,358 INFO [RS:1;7ea43b39fc24:39389 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C39389%2C1733741455809, suffix=, logDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,39389,1733741455809, archiveDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/oldWALs, maxLogs=32 2024-12-09T10:50:56,359 INFO [RS:1;7ea43b39fc24:39389 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 7ea43b39fc24%2C39389%2C1733741455809.1733741456359 2024-12-09T10:50:56,370 INFO [RS:2;7ea43b39fc24:45835 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,45835,1733741455839/7ea43b39fc24%2C45835%2C1733741455839.1733741456356 2024-12-09T10:50:56,370 INFO [RS:0;7ea43b39fc24:40291 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,40291,1733741455779/7ea43b39fc24%2C40291%2C1733741455779.1733741456356 2024-12-09T10:50:56,373 DEBUG [RS:2;7ea43b39fc24:45835 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43387:43387),(127.0.0.1/127.0.0.1:40793:40793),(127.0.0.1/127.0.0.1:37361:37361)] 2024-12-09T10:50:56,374 DEBUG [RS:0;7ea43b39fc24:40291 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43387:43387),(127.0.0.1/127.0.0.1:40793:40793),(127.0.0.1/127.0.0.1:37361:37361)] 2024-12-09T10:50:56,374 INFO [RS:1;7ea43b39fc24:39389 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,39389,1733741455809/7ea43b39fc24%2C39389%2C1733741455809.1733741456359 2024-12-09T10:50:56,376 DEBUG [RS:1;7ea43b39fc24:39389 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43387:43387),(127.0.0.1/127.0.0.1:40793:40793),(127.0.0.1/127.0.0.1:37361:37361)] 2024-12-09T10:50:56,522 DEBUG [7ea43b39fc24:46161 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-09T10:50:56,522 DEBUG [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(204): Hosts are {7ea43b39fc24=0} racks are {/default-rack=0} 2024-12-09T10:50:56,524 DEBUG [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-09T10:50:56,525 DEBUG [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-09T10:50:56,525 DEBUG [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-09T10:50:56,525 DEBUG [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-09T10:50:56,525 DEBUG [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-09T10:50:56,525 DEBUG [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-09T10:50:56,525 INFO [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-09T10:50:56,525 INFO [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-09T10:50:56,525 INFO [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-09T10:50:56,525 DEBUG [7ea43b39fc24:46161 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-09T10:50:56,525 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,527 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 7ea43b39fc24,40291,1733741455779, state=OPENING 2024-12-09T10:50:56,529 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-09T10:50:56,531 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,531 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,531 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,531 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:56,532 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:56,532 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:56,532 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-09T10:50:56,532 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:56,532 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=7ea43b39fc24,40291,1733741455779}] 2024-12-09T10:50:56,533 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:56,687 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-09T10:50:56,689 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59585, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-09T10:50:56,694 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-09T10:50:56,695 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-09T10:50:56,697 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=7ea43b39fc24%2C40291%2C1733741455779.meta, suffix=.meta, logDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,40291,1733741455779, archiveDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/oldWALs, maxLogs=32 2024-12-09T10:50:56,699 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 7ea43b39fc24%2C40291%2C1733741455779.meta.1733741456698.meta 2024-12-09T10:50:56,708 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/WALs/7ea43b39fc24,40291,1733741455779/7ea43b39fc24%2C40291%2C1733741455779.meta.1733741456698.meta 2024-12-09T10:50:56,712 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37361:37361),(127.0.0.1/127.0.0.1:40793:40793),(127.0.0.1/127.0.0.1:43387:43387)] 2024-12-09T10:50:56,715 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-09T10:50:56,715 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-09T10:50:56,716 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-09T10:50:56,716 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-09T10:50:56,716 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-09T10:50:56,716 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:56,716 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-09T10:50:56,716 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-09T10:50:56,719 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-09T10:50:56,720 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-09T10:50:56,720 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,721 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:56,721 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-09T10:50:56,722 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-09T10:50:56,722 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,723 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:56,723 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-09T10:50:56,724 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-09T10:50:56,724 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,724 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:56,724 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-09T10:50:56,725 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-09T10:50:56,726 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,726 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-09T10:50:56,726 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-09T10:50:56,727 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740 2024-12-09T10:50:56,729 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740 2024-12-09T10:50:56,730 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-09T10:50:56,730 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-09T10:50:56,731 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-09T10:50:56,733 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-09T10:50:56,734 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67044333, jitterRate=-9.615868330001831E-4}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-09T10:50:56,734 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-09T10:50:56,735 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733741456717Writing region info on filesystem at 1733741456717Initializing all the Stores at 1733741456718 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741456718Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741456719 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741456719Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733741456719Cleaning up temporary data from old regions at 1733741456730 (+11 ms)Running coprocessor post-open hooks at 1733741456734 (+4 ms)Region opened successfully at 1733741456735 (+1 ms) 2024-12-09T10:50:56,737 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733741456687 2024-12-09T10:50:56,741 DEBUG [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-09T10:50:56,741 INFO [RS_OPEN_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-09T10:50:56,742 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,743 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 7ea43b39fc24,40291,1733741455779, state=OPEN 2024-12-09T10:50:56,745 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T10:50:56,745 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T10:50:56,745 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T10:50:56,745 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-09T10:50:56,745 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:56,745 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:56,745 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:56,745 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:56,745 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-09T10:50:56,749 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-09T10:50:56,749 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=7ea43b39fc24,40291,1733741455779 in 213 msec 2024-12-09T10:50:56,753 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-09T10:50:56,753 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 632 msec 2024-12-09T10:50:56,755 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-09T10:50:56,755 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-09T10:50:56,757 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-09T10:50:56,757 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=7ea43b39fc24,40291,1733741455779, seqNum=-1] 2024-12-09T10:50:56,757 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T10:50:56,759 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:33355, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T10:50:56,768 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 711 msec 2024-12-09T10:50:56,768 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733741456768, completionTime=-1 2024-12-09T10:50:56,768 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-09T10:50:56,768 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-09T10:50:56,771 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-09T10:50:56,771 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733741516771 2024-12-09T10:50:56,771 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733741576771 2024-12-09T10:50:56,771 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-12-09T10:50:56,772 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,46161,1733741455732-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,772 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,46161,1733741455732-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,772 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,46161,1733741455732-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,772 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-7ea43b39fc24:46161, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,772 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,772 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,775 DEBUG [master/7ea43b39fc24:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-09T10:50:56,777 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.898sec 2024-12-09T10:50:56,777 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-09T10:50:56,777 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-09T10:50:56,778 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-09T10:50:56,778 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-09T10:50:56,778 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-09T10:50:56,778 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,46161,1733741455732-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-09T10:50:56,778 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,46161,1733741455732-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-09T10:50:56,781 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-09T10:50:56,781 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-09T10:50:56,781 INFO [master/7ea43b39fc24:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=7ea43b39fc24,46161,1733741455732-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-09T10:50:56,856 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3515910a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T10:50:56,856 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 7ea43b39fc24,46161,-1 for getting cluster id 2024-12-09T10:50:56,856 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-09T10:50:56,858 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'a99fdc1b-408b-4818-9d28-02d317339801' 2024-12-09T10:50:56,858 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-09T10:50:56,858 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "a99fdc1b-408b-4818-9d28-02d317339801" 2024-12-09T10:50:56,859 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@11092975, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T10:50:56,859 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [7ea43b39fc24,46161,-1] 2024-12-09T10:50:56,859 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-09T10:50:56,859 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:56,861 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45260, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-09T10:50:56,862 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6404006c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-09T10:50:56,862 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-09T10:50:56,864 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=7ea43b39fc24,40291,1733741455779, seqNum=-1] 2024-12-09T10:50:56,864 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T10:50:56,866 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:50030, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T10:50:56,868 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:56,868 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-09T10:50:56,869 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is 7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:56,870 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@607bcd5b 2024-12-09T10:50:56,870 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-09T10:50:56,871 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45274, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-09T10:50:56,872 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-09T10:50:56,874 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-09T10:50:56,876 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-09T10:50:56,876 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:56,876 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-09T10:50:56,878 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:56,878 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-09T10:50:56,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741837_1013 (size=392) 2024-12-09T10:50:56,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741837_1013 (size=392) 2024-12-09T10:50:56,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741837_1013 (size=392) 2024-12-09T10:50:56,989 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:57,199 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:57,291 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 6d163da9f338224589ef9e4ee96dc8e6, NAME => 'TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def 2024-12-09T10:50:57,301 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741838_1014 (size=51) 2024-12-09T10:50:57,301 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741838_1014 (size=51) 2024-12-09T10:50:57,301 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741838_1014 (size=51) 2024-12-09T10:50:57,302 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:57,302 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 6d163da9f338224589ef9e4ee96dc8e6, disabling compactions & flushes 2024-12-09T10:50:57,303 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:57,303 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:57,303 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. after waiting 0 ms 2024-12-09T10:50:57,303 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:57,303 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:57,303 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 6d163da9f338224589ef9e4ee96dc8e6: Waiting for close lock at 1733741457302Disabling compacts and flushes for region at 1733741457302Disabling writes for close at 1733741457303 (+1 ms)Writing region close event to WAL at 1733741457303Closed at 1733741457303 2024-12-09T10:50:57,305 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-09T10:50:57,305 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733741457305"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733741457305"}]},"ts":"1733741457305"} 2024-12-09T10:50:57,312 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-09T10:50:57,314 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-09T10:50:57,314 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733741457314"}]},"ts":"1733741457314"} 2024-12-09T10:50:57,317 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-09T10:50:57,318 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {7ea43b39fc24=0} racks are {/default-rack=0} 2024-12-09T10:50:57,319 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-09T10:50:57,319 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-09T10:50:57,319 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-09T10:50:57,319 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-09T10:50:57,319 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-09T10:50:57,319 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-09T10:50:57,319 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-09T10:50:57,319 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-09T10:50:57,319 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-09T10:50:57,319 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-09T10:50:57,319 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6d163da9f338224589ef9e4ee96dc8e6, ASSIGN}] 2024-12-09T10:50:57,321 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6d163da9f338224589ef9e4ee96dc8e6, ASSIGN 2024-12-09T10:50:57,323 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6d163da9f338224589ef9e4ee96dc8e6, ASSIGN; state=OFFLINE, location=7ea43b39fc24,45835,1733741455839; forceNewPlan=false, retain=false 2024-12-09T10:50:57,473 INFO [7ea43b39fc24:46161 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-09T10:50:57,474 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=6d163da9f338224589ef9e4ee96dc8e6, regionState=OPENING, regionLocation=7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:57,478 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6d163da9f338224589ef9e4ee96dc8e6, ASSIGN because future has completed 2024-12-09T10:50:57,478 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 6d163da9f338224589ef9e4ee96dc8e6, server=7ea43b39fc24,45835,1733741455839}] 2024-12-09T10:50:57,510 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:57,632 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-09T10:50:57,634 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51677, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-09T10:50:57,639 INFO [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:57,639 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 6d163da9f338224589ef9e4ee96dc8e6, NAME => 'TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6.', STARTKEY => '', ENDKEY => ''} 2024-12-09T10:50:57,640 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,640 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-09T10:50:57,640 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,640 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,642 INFO [StoreOpener-6d163da9f338224589ef9e4ee96dc8e6-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,644 INFO [StoreOpener-6d163da9f338224589ef9e4ee96dc8e6-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6d163da9f338224589ef9e4ee96dc8e6 columnFamilyName cf 2024-12-09T10:50:57,644 DEBUG [StoreOpener-6d163da9f338224589ef9e4ee96dc8e6-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-09T10:50:57,644 INFO [StoreOpener-6d163da9f338224589ef9e4ee96dc8e6-1 {}] regionserver.HStore(327): Store=6d163da9f338224589ef9e4ee96dc8e6/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-09T10:50:57,645 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,645 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,646 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,646 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,646 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,648 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,650 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-09T10:50:57,651 INFO [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 6d163da9f338224589ef9e4ee96dc8e6; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70799892, jitterRate=0.05500060319900513}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-09T10:50:57,651 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:57,652 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 6d163da9f338224589ef9e4ee96dc8e6: Running coprocessor pre-open hook at 1733741457640Writing region info on filesystem at 1733741457640Initializing all the Stores at 1733741457642 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733741457642Cleaning up temporary data from old regions at 1733741457646 (+4 ms)Running coprocessor post-open hooks at 1733741457651 (+5 ms)Region opened successfully at 1733741457652 (+1 ms) 2024-12-09T10:50:57,653 INFO [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6., pid=6, masterSystemTime=1733741457632 2024-12-09T10:50:57,657 DEBUG [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:57,657 INFO [RS_OPEN_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:57,658 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=6d163da9f338224589ef9e4ee96dc8e6, regionState=OPEN, openSeqNum=2, regionLocation=7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:57,661 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 6d163da9f338224589ef9e4ee96dc8e6, server=7ea43b39fc24,45835,1733741455839 because future has completed 2024-12-09T10:50:57,666 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-09T10:50:57,666 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 6d163da9f338224589ef9e4ee96dc8e6, server=7ea43b39fc24,45835,1733741455839 in 185 msec 2024-12-09T10:50:57,670 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-09T10:50:57,670 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6d163da9f338224589ef9e4ee96dc8e6, ASSIGN in 347 msec 2024-12-09T10:50:57,671 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-09T10:50:57,672 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733741457672"}]},"ts":"1733741457672"} 2024-12-09T10:50:57,675 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-09T10:50:57,676 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-09T10:50:57,679 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 804 msec 2024-12-09T10:50:57,771 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-09T10:50:57,777 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-09T10:50:57,822 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-09T10:50:57,823 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-09T10:50:58,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-09T10:50:58,020 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-09T10:50:58,020 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-09T10:50:58,020 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-09T10:50:58,027 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-09T10:50:58,027 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-09T10:50:58,027 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-09T10:50:58,030 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6., hostname=7ea43b39fc24,45835,1733741455839, seqNum=2] 2024-12-09T10:50:58,031 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-09T10:50:58,033 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46756, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-09T10:50:58,036 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestHBaseWalOnEC 2024-12-09T10:50:58,038 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-09T10:50:58,039 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T10:50:58,039 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-09T10:50:58,041 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-09T10:50:58,041 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-09T10:50:58,150 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T10:50:58,196 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45835 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-09T10:50:58,196 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:58,196 INFO [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 6d163da9f338224589ef9e4ee96dc8e6 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-09T10:50:58,216 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6/.tmp/cf/00024a68d22241489c598bd27f9c376f is 36, key is row/cf:cq/1733741458033/Put/seqid=0 2024-12-09T10:50:58,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741839_1015 (size=4787) 2024-12-09T10:50:58,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741839_1015 (size=4787) 2024-12-09T10:50:58,225 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741839_1015 (size=4787) 2024-12-09T10:50:58,226 INFO [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6/.tmp/cf/00024a68d22241489c598bd27f9c376f 2024-12-09T10:50:58,235 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6/.tmp/cf/00024a68d22241489c598bd27f9c376f as hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6/cf/00024a68d22241489c598bd27f9c376f 2024-12-09T10:50:58,243 INFO [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6/cf/00024a68d22241489c598bd27f9c376f, entries=1, sequenceid=5, filesize=4.7 K 2024-12-09T10:50:58,245 INFO [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 6d163da9f338224589ef9e4ee96dc8e6 in 49ms, sequenceid=5, compaction requested=false 2024-12-09T10:50:58,245 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 6d163da9f338224589ef9e4ee96dc8e6: 2024-12-09T10:50:58,245 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:58,246 DEBUG [RS_FLUSH_OPERATIONS-regionserver/7ea43b39fc24:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-09T10:50:58,246 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-09T10:50:58,252 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-09T10:50:58,252 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 207 msec 2024-12-09T10:50:58,255 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 217 msec 2024-12-09T10:50:58,360 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46161 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-09T10:50:58,360 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-09T10:50:58,365 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-09T10:50:58,365 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-09T10:50:58,365 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:58,365 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,365 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,365 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-09T10:50:58,365 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-09T10:50:58,365 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=25940370, stopped=false 2024-12-09T10:50:58,366 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=7ea43b39fc24,46161,1733741455732 2024-12-09T10:50:58,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:58,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:58,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:58,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-09T10:50:58,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:58,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:58,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:58,368 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-09T10:50:58,368 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:58,368 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-09T10:50:58,368 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:58,368 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,368 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:58,368 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:58,368 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:58,368 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-09T10:50:58,368 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '7ea43b39fc24,40291,1733741455779' ***** 2024-12-09T10:50:58,369 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T10:50:58,369 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '7ea43b39fc24,39389,1733741455809' ***** 2024-12-09T10:50:58,369 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T10:50:58,369 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '7ea43b39fc24,45835,1733741455839' ***** 2024-12-09T10:50:58,369 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-09T10:50:58,369 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T10:50:58,369 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T10:50:58,369 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-09T10:50:58,369 INFO [RS:0;7ea43b39fc24:40291 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T10:50:58,369 INFO [RS:0;7ea43b39fc24:40291 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T10:50:58,369 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T10:50:58,370 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(959): stopping server 7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:58,370 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T10:50:58,370 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T10:50:58,370 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-09T10:50:58,369 INFO [RS:2;7ea43b39fc24:45835 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T10:50:58,369 INFO [RS:1;7ea43b39fc24:39389 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-09T10:50:58,370 INFO [RS:2;7ea43b39fc24:45835 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T10:50:58,370 INFO [RS:0;7ea43b39fc24:40291 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;7ea43b39fc24:40291. 2024-12-09T10:50:58,370 INFO [RS:1;7ea43b39fc24:39389 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-09T10:50:58,370 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(959): stopping server 7ea43b39fc24,39389,1733741455809 2024-12-09T10:50:58,370 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(3091): Received CLOSE for 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:58,370 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T10:50:58,370 DEBUG [RS:0;7ea43b39fc24:40291 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:58,370 DEBUG [RS:0;7ea43b39fc24:40291 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,370 INFO [RS:1;7ea43b39fc24:39389 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;7ea43b39fc24:39389. 2024-12-09T10:50:58,370 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T10:50:58,370 DEBUG [RS:1;7ea43b39fc24:39389 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:58,370 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T10:50:58,370 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T10:50:58,370 DEBUG [RS:1;7ea43b39fc24:39389 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,370 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-09T10:50:58,370 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(976): stopping server 7ea43b39fc24,39389,1733741455809; all regions closed. 2024-12-09T10:50:58,371 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-09T10:50:58,371 DEBUG [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-09T10:50:58,371 DEBUG [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-09T10:50:58,371 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-09T10:50:58,371 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(959): stopping server 7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:58,371 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-09T10:50:58,371 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T10:50:58,371 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-09T10:50:58,371 INFO [RS:2;7ea43b39fc24:45835 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;7ea43b39fc24:45835. 2024-12-09T10:50:58,371 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-09T10:50:58,371 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-09T10:50:58,371 DEBUG [RS:2;7ea43b39fc24:45835 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-09T10:50:58,371 DEBUG [RS:2;7ea43b39fc24:45835 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,371 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 6d163da9f338224589ef9e4ee96dc8e6, disabling compactions & flushes 2024-12-09T10:50:58,371 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-09T10:50:58,371 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-09T10:50:58,371 INFO [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:58,371 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,371 DEBUG [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1325): Online Regions={6d163da9f338224589ef9e4ee96dc8e6=TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6.} 2024-12-09T10:50:58,371 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:58,372 DEBUG [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1351): Waiting on 6d163da9f338224589ef9e4ee96dc8e6 2024-12-09T10:50:58,372 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. after waiting 0 ms 2024-12-09T10:50:58,372 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:58,372 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,372 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,372 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,372 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,376 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741835_1011 (size=93) 2024-12-09T10:50:58,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741835_1011 (size=93) 2024-12-09T10:50:58,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741835_1011 (size=93) 2024-12-09T10:50:58,379 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/default/TestHBaseWalOnEC/6d163da9f338224589ef9e4ee96dc8e6/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-09T10:50:58,380 DEBUG [RS:1;7ea43b39fc24:39389 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/oldWALs 2024-12-09T10:50:58,380 INFO [RS:1;7ea43b39fc24:39389 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 7ea43b39fc24%2C39389%2C1733741455809:(num 1733741456359) 2024-12-09T10:50:58,380 DEBUG [RS:1;7ea43b39fc24:39389 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,380 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:58,380 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T10:50:58,380 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.ChoreService(370): Chore service for: regionserver/7ea43b39fc24:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-09T10:50:58,381 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T10:50:58,381 INFO [regionserver/7ea43b39fc24:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T10:50:58,381 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T10:50:58,381 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T10:50:58,381 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T10:50:58,381 INFO [RS:1;7ea43b39fc24:39389 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:39389 2024-12-09T10:50:58,381 INFO [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:58,381 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 6d163da9f338224589ef9e4ee96dc8e6: Waiting for close lock at 1733741458371Running coprocessor pre-close hooks at 1733741458371Disabling compacts and flushes for region at 1733741458371Disabling writes for close at 1733741458372 (+1 ms)Writing region close event to WAL at 1733741458373 (+1 ms)Running coprocessor post-close hooks at 1733741458381 (+8 ms)Closed at 1733741458381 2024-12-09T10:50:58,382 DEBUG [RS_CLOSE_REGION-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6. 2024-12-09T10:50:58,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/7ea43b39fc24,39389,1733741455809 2024-12-09T10:50:58,383 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T10:50:58,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:58,385 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [7ea43b39fc24,39389,1733741455809] 2024-12-09T10:50:58,387 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/7ea43b39fc24,39389,1733741455809 already deleted, retry=false 2024-12-09T10:50:58,387 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 7ea43b39fc24,39389,1733741455809 expired; onlineServers=2 2024-12-09T10:50:58,396 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/info/a63dab4cf6dd4bc1a44cc8eab0de8179 is 153, key is TestHBaseWalOnEC,,1733741456872.6d163da9f338224589ef9e4ee96dc8e6./info:regioninfo/1733741457658/Put/seqid=0 2024-12-09T10:50:58,397 WARN [IPC Server handler 1 on default port 34523 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-09T10:50:58,397 WARN [IPC Server handler 1 on default port 34523 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-09T10:50:58,397 WARN [IPC Server handler 1 on default port 34523 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-09T10:50:58,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741840_1016 (size=6637) 2024-12-09T10:50:58,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741840_1016 (size=6637) 2024-12-09T10:50:58,403 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/info/a63dab4cf6dd4bc1a44cc8eab0de8179 2024-12-09T10:50:58,409 INFO [regionserver/7ea43b39fc24:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:58,415 INFO [regionserver/7ea43b39fc24:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:58,417 INFO [regionserver/7ea43b39fc24:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:58,427 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/ns/b27c78a467374b8a883985101cc6c22f is 43, key is default/ns:d/1733741456760/Put/seqid=0 2024-12-09T10:50:58,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741841_1017 (size=5153) 2024-12-09T10:50:58,435 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741841_1017 (size=5153) 2024-12-09T10:50:58,435 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741841_1017 (size=5153) 2024-12-09T10:50:58,435 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/ns/b27c78a467374b8a883985101cc6c22f 2024-12-09T10:50:58,460 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/table/c74510d35f964fc389c4e9a8a4f885a0 is 52, key is TestHBaseWalOnEC/table:state/1733741457672/Put/seqid=0 2024-12-09T10:50:58,461 WARN [IPC Server handler 3 on default port 34523 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-09T10:50:58,461 WARN [IPC Server handler 3 on default port 34523 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-09T10:50:58,461 WARN [IPC Server handler 3 on default port 34523 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-09T10:50:58,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741842_1018 (size=5249) 2024-12-09T10:50:58,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741842_1018 (size=5249) 2024-12-09T10:50:58,467 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/table/c74510d35f964fc389c4e9a8a4f885a0 2024-12-09T10:50:58,474 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/info/a63dab4cf6dd4bc1a44cc8eab0de8179 as hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/info/a63dab4cf6dd4bc1a44cc8eab0de8179 2024-12-09T10:50:58,482 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/info/a63dab4cf6dd4bc1a44cc8eab0de8179, entries=10, sequenceid=11, filesize=6.5 K 2024-12-09T10:50:58,483 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/ns/b27c78a467374b8a883985101cc6c22f as hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/ns/b27c78a467374b8a883985101cc6c22f 2024-12-09T10:50:58,485 INFO [RS:1;7ea43b39fc24:39389 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T10:50:58,485 INFO [RS:1;7ea43b39fc24:39389 {}] regionserver.HRegionServer(1031): Exiting; stopping=7ea43b39fc24,39389,1733741455809; zookeeper connection closed. 2024-12-09T10:50:58,486 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:58,486 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@28eec86 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@28eec86 2024-12-09T10:50:58,486 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39389-0x100bd5f48660002, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:58,492 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/ns/b27c78a467374b8a883985101cc6c22f, entries=2, sequenceid=11, filesize=5.0 K 2024-12-09T10:50:58,493 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/.tmp/table/c74510d35f964fc389c4e9a8a4f885a0 as hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/table/c74510d35f964fc389c4e9a8a4f885a0 2024-12-09T10:50:58,500 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/table/c74510d35f964fc389c4e9a8a4f885a0, entries=2, sequenceid=11, filesize=5.1 K 2024-12-09T10:50:58,502 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 131ms, sequenceid=11, compaction requested=false 2024-12-09T10:50:58,508 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-09T10:50:58,508 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-09T10:50:58,509 INFO [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-09T10:50:58,509 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733741458371Running coprocessor pre-close hooks at 1733741458371Disabling compacts and flushes for region at 1733741458371Disabling writes for close at 1733741458371Obtaining lock to block concurrent updates at 1733741458371Preparing flush snapshotting stores in 1588230740 at 1733741458371Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733741458372 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733741458375 (+3 ms)Flushing 1588230740/info: creating writer at 1733741458375Flushing 1588230740/info: appending metadata at 1733741458395 (+20 ms)Flushing 1588230740/info: closing flushed file at 1733741458395Flushing 1588230740/ns: creating writer at 1733741458411 (+16 ms)Flushing 1588230740/ns: appending metadata at 1733741458427 (+16 ms)Flushing 1588230740/ns: closing flushed file at 1733741458427Flushing 1588230740/table: creating writer at 1733741458443 (+16 ms)Flushing 1588230740/table: appending metadata at 1733741458459 (+16 ms)Flushing 1588230740/table: closing flushed file at 1733741458459Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4b9845b8: reopening flushed file at 1733741458473 (+14 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5d98a628: reopening flushed file at 1733741458482 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@74a68a7: reopening flushed file at 1733741458492 (+10 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 131ms, sequenceid=11, compaction requested=false at 1733741458502 (+10 ms)Writing region close event to WAL at 1733741458503 (+1 ms)Running coprocessor post-close hooks at 1733741458508 (+5 ms)Closed at 1733741458509 (+1 ms) 2024-12-09T10:50:58,509 DEBUG [RS_CLOSE_META-regionserver/7ea43b39fc24:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-09T10:50:58,571 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(976): stopping server 7ea43b39fc24,40291,1733741455779; all regions closed. 2024-12-09T10:50:58,572 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(976): stopping server 7ea43b39fc24,45835,1733741455839; all regions closed. 2024-12-09T10:50:58,572 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,572 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,572 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,572 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,572 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,572 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,572 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,573 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,573 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,573 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741836_1012 (size=2751) 2024-12-09T10:50:58,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741834_1010 (size=1298) 2024-12-09T10:50:58,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741836_1012 (size=2751) 2024-12-09T10:50:58,578 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741834_1010 (size=1298) 2024-12-09T10:50:58,578 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741834_1010 (size=1298) 2024-12-09T10:50:58,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741836_1012 (size=2751) 2024-12-09T10:50:58,580 DEBUG [RS:2;7ea43b39fc24:45835 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/oldWALs 2024-12-09T10:50:58,580 INFO [RS:2;7ea43b39fc24:45835 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 7ea43b39fc24%2C45835%2C1733741455839:(num 1733741456356) 2024-12-09T10:50:58,580 DEBUG [RS:2;7ea43b39fc24:45835 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,580 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:58,580 DEBUG [RS:0;7ea43b39fc24:40291 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/oldWALs 2024-12-09T10:50:58,580 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T10:50:58,580 INFO [RS:0;7ea43b39fc24:40291 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 7ea43b39fc24%2C40291%2C1733741455779.meta:.meta(num 1733741456698) 2024-12-09T10:50:58,580 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.ChoreService(370): Chore service for: regionserver/7ea43b39fc24:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-09T10:50:58,581 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-09T10:50:58,581 INFO [regionserver/7ea43b39fc24:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T10:50:58,581 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-09T10:50:58,581 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-09T10:50:58,581 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T10:50:58,581 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,581 INFO [RS:2;7ea43b39fc24:45835 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:45835 2024-12-09T10:50:58,581 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,581 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,582 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,582 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,584 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/7ea43b39fc24,45835,1733741455839 2024-12-09T10:50:58,584 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T10:50:58,584 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:58,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741833_1009 (size=93) 2024-12-09T10:50:58,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741833_1009 (size=93) 2024-12-09T10:50:58,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741833_1009 (size=93) 2024-12-09T10:50:58,586 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [7ea43b39fc24,45835,1733741455839] 2024-12-09T10:50:58,587 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/7ea43b39fc24,45835,1733741455839 already deleted, retry=false 2024-12-09T10:50:58,587 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 7ea43b39fc24,45835,1733741455839 expired; onlineServers=1 2024-12-09T10:50:58,589 DEBUG [RS:0;7ea43b39fc24:40291 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/oldWALs 2024-12-09T10:50:58,589 INFO [RS:0;7ea43b39fc24:40291 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 7ea43b39fc24%2C40291%2C1733741455779:(num 1733741456356) 2024-12-09T10:50:58,589 DEBUG [RS:0;7ea43b39fc24:40291 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-09T10:50:58,589 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.LeaseManager(133): Closed leases 2024-12-09T10:50:58,589 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T10:50:58,589 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.ChoreService(370): Chore service for: regionserver/7ea43b39fc24:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-09T10:50:58,589 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T10:50:58,590 INFO [regionserver/7ea43b39fc24:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T10:50:58,590 INFO [RS:0;7ea43b39fc24:40291 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:40291 2024-12-09T10:50:58,592 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/7ea43b39fc24,40291,1733741455779 2024-12-09T10:50:58,592 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T10:50:58,593 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-09T10:50:58,594 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [7ea43b39fc24,40291,1733741455779] 2024-12-09T10:50:58,596 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/7ea43b39fc24,40291,1733741455779 already deleted, retry=false 2024-12-09T10:50:58,596 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 7ea43b39fc24,40291,1733741455779 expired; onlineServers=0 2024-12-09T10:50:58,596 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '7ea43b39fc24,46161,1733741455732' ***** 2024-12-09T10:50:58,596 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-09T10:50:58,596 INFO [M:0;7ea43b39fc24:46161 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-09T10:50:58,596 INFO [M:0;7ea43b39fc24:46161 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-09T10:50:58,596 DEBUG [M:0;7ea43b39fc24:46161 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-09T10:50:58,596 DEBUG [M:0;7ea43b39fc24:46161 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-09T10:50:58,596 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.large.0-1733741456069 {}] cleaner.HFileCleaner(306): Exit Thread[master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.large.0-1733741456069,5,FailOnTimeoutGroup] 2024-12-09T10:50:58,596 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-09T10:50:58,596 DEBUG [master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.small.0-1733741456069 {}] cleaner.HFileCleaner(306): Exit Thread[master/7ea43b39fc24:0:becomeActiveMaster-HFileCleaner.small.0-1733741456069,5,FailOnTimeoutGroup] 2024-12-09T10:50:58,597 INFO [M:0;7ea43b39fc24:46161 {}] hbase.ChoreService(370): Chore service for: master/7ea43b39fc24:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-09T10:50:58,597 INFO [M:0;7ea43b39fc24:46161 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-09T10:50:58,597 DEBUG [M:0;7ea43b39fc24:46161 {}] master.HMaster(1795): Stopping service threads 2024-12-09T10:50:58,597 INFO [M:0;7ea43b39fc24:46161 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-09T10:50:58,597 INFO [M:0;7ea43b39fc24:46161 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-09T10:50:58,597 INFO [M:0;7ea43b39fc24:46161 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-09T10:50:58,597 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-09T10:50:58,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-09T10:50:58,599 DEBUG [M:0;7ea43b39fc24:46161 {}] zookeeper.ZKUtil(347): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-09T10:50:58,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-09T10:50:58,599 WARN [M:0;7ea43b39fc24:46161 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-09T10:50:58,599 INFO [M:0;7ea43b39fc24:46161 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/.lastflushedseqids 2024-12-09T10:50:58,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741843_1019 (size=127) 2024-12-09T10:50:58,608 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741843_1019 (size=127) 2024-12-09T10:50:58,608 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741843_1019 (size=127) 2024-12-09T10:50:58,608 INFO [M:0;7ea43b39fc24:46161 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-09T10:50:58,609 INFO [M:0;7ea43b39fc24:46161 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-09T10:50:58,609 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-09T10:50:58,609 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:58,609 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:58,609 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-09T10:50:58,609 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:58,609 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.85 KB heapSize=34.13 KB 2024-12-09T10:50:58,628 DEBUG [M:0;7ea43b39fc24:46161 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/42734fc1d43d40248849b7ec4dc54f64 is 82, key is hbase:meta,,1/info:regioninfo/1733741456742/Put/seqid=0 2024-12-09T10:50:58,636 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741844_1020 (size=5672) 2024-12-09T10:50:58,636 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741844_1020 (size=5672) 2024-12-09T10:50:58,637 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741844_1020 (size=5672) 2024-12-09T10:50:58,637 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/42734fc1d43d40248849b7ec4dc54f64 2024-12-09T10:50:58,668 DEBUG [M:0;7ea43b39fc24:46161 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/b2b5883dc80e40faa7238edf1d2bf9d3 is 749, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733741457678/Put/seqid=0 2024-12-09T10:50:58,676 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741845_1021 (size=6441) 2024-12-09T10:50:58,677 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741845_1021 (size=6441) 2024-12-09T10:50:58,677 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741845_1021 (size=6441) 2024-12-09T10:50:58,678 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.17 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/b2b5883dc80e40faa7238edf1d2bf9d3 2024-12-09T10:50:58,686 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:58,686 INFO [RS:2;7ea43b39fc24:45835 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T10:50:58,686 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45835-0x100bd5f48660003, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:58,686 INFO [RS:2;7ea43b39fc24:45835 {}] regionserver.HRegionServer(1031): Exiting; stopping=7ea43b39fc24,45835,1733741455839; zookeeper connection closed. 2024-12-09T10:50:58,690 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4704ee07 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4704ee07 2024-12-09T10:50:58,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:58,695 INFO [RS:0;7ea43b39fc24:40291 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T10:50:58,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40291-0x100bd5f48660001, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:58,695 INFO [RS:0;7ea43b39fc24:40291 {}] regionserver.HRegionServer(1031): Exiting; stopping=7ea43b39fc24,40291,1733741455779; zookeeper connection closed. 2024-12-09T10:50:58,695 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@8afa8d {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@8afa8d 2024-12-09T10:50:58,695 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-09T10:50:58,704 DEBUG [M:0;7ea43b39fc24:46161 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/bb98c75c3a5143c6b1932cf3168bef44 is 69, key is 7ea43b39fc24,39389,1733741455809/rs:state/1733741456180/Put/seqid=0 2024-12-09T10:50:58,713 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741846_1022 (size=5294) 2024-12-09T10:50:58,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741846_1022 (size=5294) 2024-12-09T10:50:58,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741846_1022 (size=5294) 2024-12-09T10:50:58,715 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/bb98c75c3a5143c6b1932cf3168bef44 2024-12-09T10:50:58,723 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/42734fc1d43d40248849b7ec4dc54f64 as hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/42734fc1d43d40248849b7ec4dc54f64 2024-12-09T10:50:58,732 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/42734fc1d43d40248849b7ec4dc54f64, entries=8, sequenceid=72, filesize=5.5 K 2024-12-09T10:50:58,734 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/b2b5883dc80e40faa7238edf1d2bf9d3 as hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/b2b5883dc80e40faa7238edf1d2bf9d3 2024-12-09T10:50:58,742 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/b2b5883dc80e40faa7238edf1d2bf9d3, entries=8, sequenceid=72, filesize=6.3 K 2024-12-09T10:50:58,744 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/bb98c75c3a5143c6b1932cf3168bef44 as hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/bb98c75c3a5143c6b1932cf3168bef44 2024-12-09T10:50:58,751 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34523/user/jenkins/test-data/82b83f5a-23e8-7c47-b62e-2b039b023def/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/bb98c75c3a5143c6b1932cf3168bef44, entries=3, sequenceid=72, filesize=5.2 K 2024-12-09T10:50:58,753 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 144ms, sequenceid=72, compaction requested=false 2024-12-09T10:50:58,754 INFO [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-09T10:50:58,755 DEBUG [M:0;7ea43b39fc24:46161 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733741458609Disabling compacts and flushes for region at 1733741458609Disabling writes for close at 1733741458609Obtaining lock to block concurrent updates at 1733741458609Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733741458609Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27492, getHeapSize=34888, getOffHeapSize=0, getCellsCount=85 at 1733741458610 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733741458610Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733741458611 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733741458628 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733741458628Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733741458644 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733741458668 (+24 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733741458668Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733741458685 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733741458704 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733741458704Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4109ade0: reopening flushed file at 1733741458722 (+18 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@26469c2c: reopening flushed file at 1733741458733 (+11 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@d3071e3: reopening flushed file at 1733741458743 (+10 ms)Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 144ms, sequenceid=72, compaction requested=false at 1733741458753 (+10 ms)Writing region close event to WAL at 1733741458754 (+1 ms)Closed at 1733741458754 2024-12-09T10:50:58,755 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,755 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,755 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,755 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,755 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-09T10:50:58,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36897 is added to blk_1073741830_1006 (size=32695) 2024-12-09T10:50:58,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44245 is added to blk_1073741830_1006 (size=32695) 2024-12-09T10:50:58,759 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36347 is added to blk_1073741830_1006 (size=32695) 2024-12-09T10:50:58,760 INFO [M:0;7ea43b39fc24:46161 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-09T10:50:58,760 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-09T10:50:58,760 INFO [M:0;7ea43b39fc24:46161 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:46161 2024-12-09T10:50:58,760 INFO [M:0;7ea43b39fc24:46161 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-09T10:50:58,862 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:58,862 INFO [M:0;7ea43b39fc24:46161 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-09T10:50:58,862 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46161-0x100bd5f48660000, quorum=127.0.0.1:59671, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-09T10:50:58,865 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@472e7194{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:58,865 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7ba9a3d5{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T10:50:58,865 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T10:50:58,865 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@19f40ccf{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T10:50:58,866 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@739551bc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir/,STOPPED} 2024-12-09T10:50:58,867 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T10:50:58,867 WARN [BP-1261905413-172.17.0.3-1733741454874 heartbeating to localhost/127.0.0.1:34523 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T10:50:58,867 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T10:50:58,867 WARN [BP-1261905413-172.17.0.3-1733741454874 heartbeating to localhost/127.0.0.1:34523 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1261905413-172.17.0.3-1733741454874 (Datanode Uuid f7633a3e-9018-4036-a2e6-39607bb150ce) service to localhost/127.0.0.1:34523 2024-12-09T10:50:58,868 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data5/current/BP-1261905413-172.17.0.3-1733741454874 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:58,869 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data6/current/BP-1261905413-172.17.0.3-1733741454874 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:58,869 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T10:50:58,871 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@18a67058{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:58,871 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@478131f{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T10:50:58,871 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T10:50:58,871 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1fee469f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T10:50:58,871 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6f8ad177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir/,STOPPED} 2024-12-09T10:50:58,873 WARN [BP-1261905413-172.17.0.3-1733741454874 heartbeating to localhost/127.0.0.1:34523 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T10:50:58,873 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T10:50:58,873 WARN [BP-1261905413-172.17.0.3-1733741454874 heartbeating to localhost/127.0.0.1:34523 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1261905413-172.17.0.3-1733741454874 (Datanode Uuid 6dc06157-63cc-42b2-b5b3-4e205a2b6465) service to localhost/127.0.0.1:34523 2024-12-09T10:50:58,873 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T10:50:58,873 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data3/current/BP-1261905413-172.17.0.3-1733741454874 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:58,873 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data4/current/BP-1261905413-172.17.0.3-1733741454874 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:58,874 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T10:50:58,876 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@38845bbf{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-09T10:50:58,876 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@24815bd8{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T10:50:58,876 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T10:50:58,876 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2591ff9a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T10:50:58,876 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7c2dd4e6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir/,STOPPED} 2024-12-09T10:50:58,877 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-09T10:50:58,877 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-09T10:50:58,877 WARN [BP-1261905413-172.17.0.3-1733741454874 heartbeating to localhost/127.0.0.1:34523 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-09T10:50:58,878 WARN [BP-1261905413-172.17.0.3-1733741454874 heartbeating to localhost/127.0.0.1:34523 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1261905413-172.17.0.3-1733741454874 (Datanode Uuid 17ec257c-95f6-41e5-98a3-f3eed91d4d86) service to localhost/127.0.0.1:34523 2024-12-09T10:50:58,878 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data1/current/BP-1261905413-172.17.0.3-1733741454874 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:58,878 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/cluster_73164008-5ff7-62d4-fe08-5b53b71886b1/data/data2/current/BP-1261905413-172.17.0.3-1733741454874 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-09T10:50:58,879 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-09T10:50:58,889 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@46039787{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-09T10:50:58,889 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7326bb42{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-09T10:50:58,890 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-09T10:50:58,890 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@71b7cabb{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-09T10:50:58,890 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@715f09c8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4be804b0-ab79-3151-90f1-4f16f1a0829b/hadoop.log.dir/,STOPPED} 2024-12-09T10:50:58,898 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-09T10:50:58,922 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-09T10:50:58,930 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=152 (was 92) - Thread LEAK? -, OpenFileDescriptor=519 (was 441) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=191 (was 199), ProcessCount=11 (was 11), AvailableMemoryMB=7594 (was 7769)