2024-12-08 03:46:02,157 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-08 03:46:02,170 main DEBUG Took 0.011615 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-08 03:46:02,171 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-08 03:46:02,171 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-08 03:46:02,172 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-08 03:46:02,174 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,189 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-08 03:46:02,205 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,207 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,207 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,208 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,209 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,209 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,210 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,211 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,211 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,212 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,213 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,213 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,214 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,215 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,215 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,216 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,216 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,216 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,217 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,217 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,218 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,218 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,219 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,219 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-08 03:46:02,220 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,220 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-08 03:46:02,222 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-08 03:46:02,223 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-08 03:46:02,225 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-08 03:46:02,225 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-08 03:46:02,227 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-08 03:46:02,228 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-08 03:46:02,237 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-08 03:46:02,240 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-08 03:46:02,242 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-08 03:46:02,243 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-08 03:46:02,244 main DEBUG createAppenders(={Console}) 2024-12-08 03:46:02,245 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-08 03:46:02,245 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-08 03:46:02,245 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-08 03:46:02,246 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-08 03:46:02,246 main DEBUG OutputStream closed 2024-12-08 03:46:02,247 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-08 03:46:02,247 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-08 03:46:02,248 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-08 03:46:02,327 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-08 03:46:02,330 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-08 03:46:02,332 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-08 03:46:02,333 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-08 03:46:02,334 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-08 03:46:02,334 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-08 03:46:02,335 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-08 03:46:02,335 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-08 03:46:02,336 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-08 03:46:02,336 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-08 03:46:02,337 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-08 03:46:02,337 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-08 03:46:02,338 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-08 03:46:02,338 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-08 03:46:02,338 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-08 03:46:02,338 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-08 03:46:02,339 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-08 03:46:02,340 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-08 03:46:02,341 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-08 03:46:02,342 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-08 03:46:02,342 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-08 03:46:02,343 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-08T03:46:02,358 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-08 03:46:02,361 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-08 03:46:02,362 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-08T03:46:02,586 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9 2024-12-08T03:46:02,612 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797, deleteOnExit=true 2024-12-08T03:46:02,613 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/test.cache.data in system properties and HBase conf 2024-12-08T03:46:02,613 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.tmp.dir in system properties and HBase conf 2024-12-08T03:46:02,614 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir in system properties and HBase conf 2024-12-08T03:46:02,614 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-08T03:46:02,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-08T03:46:02,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-08T03:46:02,687 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-08T03:46:02,765 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-08T03:46:02,769 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-08T03:46:02,769 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-08T03:46:02,770 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-08T03:46:02,770 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-08T03:46:02,770 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-08T03:46:02,771 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-08T03:46:02,771 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-08T03:46:02,772 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-08T03:46:02,772 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-08T03:46:02,772 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/nfs.dump.dir in system properties and HBase conf 2024-12-08T03:46:02,773 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/java.io.tmpdir in system properties and HBase conf 2024-12-08T03:46:02,773 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-08T03:46:02,773 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-08T03:46:02,773 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-08T03:46:03,661 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-08T03:46:03,723 INFO [Time-limited test {}] log.Log(170): Logging initialized @2167ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-08T03:46:03,788 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:03,844 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-08T03:46:03,863 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-08T03:46:03,863 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-08T03:46:03,865 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-08T03:46:03,876 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:03,879 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4395d44b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir/,AVAILABLE} 2024-12-08T03:46:03,879 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4f93dd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-08T03:46:04,039 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@58dbf239{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/java.io.tmpdir/jetty-localhost-38995-hadoop-hdfs-3_4_1-tests_jar-_-any-385105674523279713/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-08T03:46:04,046 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@13e2962d{HTTP/1.1, (http/1.1)}{localhost:38995} 2024-12-08T03:46:04,047 INFO [Time-limited test {}] server.Server(415): Started @2492ms 2024-12-08T03:46:04,531 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:04,537 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-08T03:46:04,538 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-08T03:46:04,538 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-08T03:46:04,538 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-08T03:46:04,539 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@431e53b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir/,AVAILABLE} 2024-12-08T03:46:04,540 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4dc262e0{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-08T03:46:04,632 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@65462677{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/java.io.tmpdir/jetty-localhost-41031-hadoop-hdfs-3_4_1-tests_jar-_-any-10340909927621171735/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:04,632 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@383014b{HTTP/1.1, (http/1.1)}{localhost:41031} 2024-12-08T03:46:04,633 INFO [Time-limited test {}] server.Server(415): Started @3078ms 2024-12-08T03:46:04,677 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-08T03:46:04,774 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:04,778 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-08T03:46:04,781 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-08T03:46:04,781 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-08T03:46:04,781 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-08T03:46:04,782 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@444b27d4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir/,AVAILABLE} 2024-12-08T03:46:04,783 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6af5a446{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-08T03:46:04,885 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@513cab2c{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/java.io.tmpdir/jetty-localhost-35827-hadoop-hdfs-3_4_1-tests_jar-_-any-13499440905704294329/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:04,885 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@29a123ec{HTTP/1.1, (http/1.1)}{localhost:35827} 2024-12-08T03:46:04,886 INFO [Time-limited test {}] server.Server(415): Started @3331ms 2024-12-08T03:46:04,888 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-08T03:46:04,922 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:04,926 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-08T03:46:04,927 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-08T03:46:04,927 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-08T03:46:04,927 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-08T03:46:04,928 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@35e2f174{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir/,AVAILABLE} 2024-12-08T03:46:04,929 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@343b36c2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-08T03:46:05,023 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@653e6301{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/java.io.tmpdir/jetty-localhost-45919-hadoop-hdfs-3_4_1-tests_jar-_-any-1498769384077788813/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:05,024 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@404caff2{HTTP/1.1, (http/1.1)}{localhost:45919} 2024-12-08T03:46:05,024 INFO [Time-limited test {}] server.Server(415): Started @3470ms 2024-12-08T03:46:05,026 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-08T03:46:05,920 WARN [Thread-122 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data1/current/BP-1980098496-172.17.0.2-1733629563223/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:05,920 WARN [Thread-125 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data4/current/BP-1980098496-172.17.0.2-1733629563223/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:05,920 WARN [Thread-123 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data2/current/BP-1980098496-172.17.0.2-1733629563223/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:05,920 WARN [Thread-124 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data3/current/BP-1980098496-172.17.0.2-1733629563223/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:05,953 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-08T03:46:05,953 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-08T03:46:05,965 WARN [Thread-142 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data5/current/BP-1980098496-172.17.0.2-1733629563223/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:05,966 WARN [Thread-143 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data6/current/BP-1980098496-172.17.0.2-1733629563223/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:05,991 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-08T03:46:06,001 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x23bd4da8b3e5a82a with lease ID 0x60c6e5f2781acc7b: Processing first storage report for DS-f59ef6c5-2a7a-47c5-8382-e2da01c985c3 from datanode DatanodeRegistration(127.0.0.1:35923, datanodeUuid=8930704c-b215-4d25-9b2f-8d95d2ea3622, infoPort=39763, infoSecurePort=0, ipcPort=38165, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223) 2024-12-08T03:46:06,002 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x23bd4da8b3e5a82a with lease ID 0x60c6e5f2781acc7b: from storage DS-f59ef6c5-2a7a-47c5-8382-e2da01c985c3 node DatanodeRegistration(127.0.0.1:35923, datanodeUuid=8930704c-b215-4d25-9b2f-8d95d2ea3622, infoPort=39763, infoSecurePort=0, ipcPort=38165, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-08T03:46:06,003 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xcd89e37350bf06e7 with lease ID 0x60c6e5f2781acc7a: Processing first storage report for DS-58eda793-1940-40cf-a8ce-438598725779 from datanode DatanodeRegistration(127.0.0.1:38097, datanodeUuid=f3d0217e-5dbb-44f3-96a2-2e3d2dcf7210, infoPort=44921, infoSecurePort=0, ipcPort=33407, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223) 2024-12-08T03:46:06,003 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xcd89e37350bf06e7 with lease ID 0x60c6e5f2781acc7a: from storage DS-58eda793-1940-40cf-a8ce-438598725779 node DatanodeRegistration(127.0.0.1:38097, datanodeUuid=f3d0217e-5dbb-44f3-96a2-2e3d2dcf7210, infoPort=44921, infoSecurePort=0, ipcPort=33407, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:06,003 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x35f286902fb70561 with lease ID 0x60c6e5f2781acc79: Processing first storage report for DS-6a00e5d0-2bfe-4ecf-871f-2bff8252b118 from datanode DatanodeRegistration(127.0.0.1:45143, datanodeUuid=68cb5d41-74fa-475e-9d87-aa1ee7d5825d, infoPort=38593, infoSecurePort=0, ipcPort=43763, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223) 2024-12-08T03:46:06,003 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x35f286902fb70561 with lease ID 0x60c6e5f2781acc79: from storage DS-6a00e5d0-2bfe-4ecf-871f-2bff8252b118 node DatanodeRegistration(127.0.0.1:45143, datanodeUuid=68cb5d41-74fa-475e-9d87-aa1ee7d5825d, infoPort=38593, infoSecurePort=0, ipcPort=43763, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:06,004 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x23bd4da8b3e5a82a with lease ID 0x60c6e5f2781acc7b: Processing first storage report for DS-1410c52e-5a70-4181-b1bd-93334844a5c7 from datanode DatanodeRegistration(127.0.0.1:35923, datanodeUuid=8930704c-b215-4d25-9b2f-8d95d2ea3622, infoPort=39763, infoSecurePort=0, ipcPort=38165, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223) 2024-12-08T03:46:06,004 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x23bd4da8b3e5a82a with lease ID 0x60c6e5f2781acc7b: from storage DS-1410c52e-5a70-4181-b1bd-93334844a5c7 node DatanodeRegistration(127.0.0.1:35923, datanodeUuid=8930704c-b215-4d25-9b2f-8d95d2ea3622, infoPort=39763, infoSecurePort=0, ipcPort=38165, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:06,004 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xcd89e37350bf06e7 with lease ID 0x60c6e5f2781acc7a: Processing first storage report for DS-4cceaff2-108d-40e3-97dc-eab39292914e from datanode DatanodeRegistration(127.0.0.1:38097, datanodeUuid=f3d0217e-5dbb-44f3-96a2-2e3d2dcf7210, infoPort=44921, infoSecurePort=0, ipcPort=33407, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223) 2024-12-08T03:46:06,004 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xcd89e37350bf06e7 with lease ID 0x60c6e5f2781acc7a: from storage DS-4cceaff2-108d-40e3-97dc-eab39292914e node DatanodeRegistration(127.0.0.1:38097, datanodeUuid=f3d0217e-5dbb-44f3-96a2-2e3d2dcf7210, infoPort=44921, infoSecurePort=0, ipcPort=33407, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:06,004 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x35f286902fb70561 with lease ID 0x60c6e5f2781acc79: Processing first storage report for DS-c9bd1eb6-ddf7-4cf0-88a9-83c26ef23b0a from datanode DatanodeRegistration(127.0.0.1:45143, datanodeUuid=68cb5d41-74fa-475e-9d87-aa1ee7d5825d, infoPort=38593, infoSecurePort=0, ipcPort=43763, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223) 2024-12-08T03:46:06,004 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x35f286902fb70561 with lease ID 0x60c6e5f2781acc79: from storage DS-c9bd1eb6-ddf7-4cf0-88a9-83c26ef23b0a node DatanodeRegistration(127.0.0.1:45143, datanodeUuid=68cb5d41-74fa-475e-9d87-aa1ee7d5825d, infoPort=38593, infoSecurePort=0, ipcPort=43763, storageInfo=lv=-57;cid=testClusterID;nsid=842834686;c=1733629563223), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:06,017 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9 2024-12-08T03:46:06,080 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-08T03:46:06,127 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=160, OpenFileDescriptor=393, MaxFileDescriptor=1048576, SystemLoadAverage=136, ProcessCount=11, AvailableMemoryMB=17806 2024-12-08T03:46:06,129 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-08T03:46:06,139 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-08T03:46:06,229 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/zookeeper_0, clientPort=56008, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-08T03:46:06,238 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=56008 2024-12-08T03:46:06,259 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:06,263 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:06,372 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:06,372 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:06,425 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:46648 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:38097:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46648 dst: /127.0.0.1:38097 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:06,452 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-08T03:46:06,848 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:06,859 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791 with version=8 2024-12-08T03:46:06,859 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/hbase-staging 2024-12-08T03:46:06,941 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-08T03:46:07,181 INFO [Time-limited test {}] client.ConnectionUtils(128): master/32df4b4bb439:0 server-side Connection retries=45 2024-12-08T03:46:07,190 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,190 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,194 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-08T03:46:07,195 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,195 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-08T03:46:07,313 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-08T03:46:07,367 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-08T03:46:07,374 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-08T03:46:07,377 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-08T03:46:07,399 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 19598 (auto-detected) 2024-12-08T03:46:07,400 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-08T03:46:07,416 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:37833 2024-12-08T03:46:07,434 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:37833 connecting to ZooKeeper ensemble=127.0.0.1:56008 2024-12-08T03:46:07,530 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:378330x0, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-08T03:46:07,535 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:37833-0x10003959ba10000 connected 2024-12-08T03:46:07,620 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:07,624 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:07,635 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:07,639 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791, hbase.cluster.distributed=false 2024-12-08T03:46:07,658 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-08T03:46:07,662 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37833 2024-12-08T03:46:07,662 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37833 2024-12-08T03:46:07,663 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37833 2024-12-08T03:46:07,665 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37833 2024-12-08T03:46:07,665 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37833 2024-12-08T03:46:07,756 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32df4b4bb439:0 server-side Connection retries=45 2024-12-08T03:46:07,758 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,758 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,758 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-08T03:46:07,758 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,759 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-08T03:46:07,761 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-08T03:46:07,763 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-08T03:46:07,764 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:38719 2024-12-08T03:46:07,767 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:38719 connecting to ZooKeeper ensemble=127.0.0.1:56008 2024-12-08T03:46:07,768 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:07,772 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:07,791 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:387190x0, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-08T03:46:07,792 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:38719-0x10003959ba10001 connected 2024-12-08T03:46:07,792 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:07,796 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-08T03:46:07,803 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-08T03:46:07,806 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-08T03:46:07,810 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-08T03:46:07,811 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38719 2024-12-08T03:46:07,812 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38719 2024-12-08T03:46:07,813 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38719 2024-12-08T03:46:07,813 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38719 2024-12-08T03:46:07,814 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38719 2024-12-08T03:46:07,830 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32df4b4bb439:0 server-side Connection retries=45 2024-12-08T03:46:07,830 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,830 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,831 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-08T03:46:07,831 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,831 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-08T03:46:07,831 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-08T03:46:07,832 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-08T03:46:07,832 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:40985 2024-12-08T03:46:07,834 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:40985 connecting to ZooKeeper ensemble=127.0.0.1:56008 2024-12-08T03:46:07,835 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:07,838 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:07,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:409850x0, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-08T03:46:07,876 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:40985-0x10003959ba10002 connected 2024-12-08T03:46:07,876 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:07,877 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-08T03:46:07,878 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-08T03:46:07,880 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-08T03:46:07,882 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-08T03:46:07,883 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40985 2024-12-08T03:46:07,884 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40985 2024-12-08T03:46:07,885 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40985 2024-12-08T03:46:07,885 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40985 2024-12-08T03:46:07,886 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40985 2024-12-08T03:46:07,903 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32df4b4bb439:0 server-side Connection retries=45 2024-12-08T03:46:07,904 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,904 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,904 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-08T03:46:07,904 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:07,905 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-08T03:46:07,905 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-08T03:46:07,905 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-08T03:46:07,906 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36577 2024-12-08T03:46:07,908 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36577 connecting to ZooKeeper ensemble=127.0.0.1:56008 2024-12-08T03:46:07,910 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:07,914 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:07,949 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:365770x0, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-08T03:46:07,949 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36577-0x10003959ba10003 connected 2024-12-08T03:46:07,950 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:07,950 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-08T03:46:07,951 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-08T03:46:07,952 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-08T03:46:07,954 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-08T03:46:07,955 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36577 2024-12-08T03:46:07,955 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36577 2024-12-08T03:46:07,956 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36577 2024-12-08T03:46:07,956 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36577 2024-12-08T03:46:07,957 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36577 2024-12-08T03:46:07,970 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;32df4b4bb439:37833 2024-12-08T03:46:07,971 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/32df4b4bb439,37833,1733629567036 2024-12-08T03:46:07,984 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:07,984 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:07,984 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:07,984 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:07,986 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/32df4b4bb439,37833,1733629567036 2024-12-08T03:46:08,016 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,016 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-08T03:46:08,016 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-08T03:46:08,016 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-08T03:46:08,017 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,017 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,017 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,018 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-08T03:46:08,019 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/32df4b4bb439,37833,1733629567036 from backup master directory 2024-12-08T03:46:08,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:08,025 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:08,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:08,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/32df4b4bb439,37833,1733629567036 2024-12-08T03:46:08,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:08,026 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-08T03:46:08,027 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=32df4b4bb439,37833,1733629567036 2024-12-08T03:46:08,029 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-08T03:46:08,030 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-08T03:46:08,081 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/hbase.id] with ID: 0a3cdf63-875c-4db4-907e-913e10d7eee1 2024-12-08T03:46:08,081 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/.tmp/hbase.id 2024-12-08T03:46:08,087 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:08,088 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:08,091 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:33150 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:35923:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33150 dst: /127.0.0.1:35923 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:08,096 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-08T03:46:08,097 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:08,097 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/.tmp/hbase.id]:[hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/hbase.id] 2024-12-08T03:46:08,141 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:08,146 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-08T03:46:08,164 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 17ms. 2024-12-08T03:46:08,198 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,198 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,198 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,198 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,212 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:08,212 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:08,215 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:46674 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:38097:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46674 dst: /127.0.0.1:38097 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:08,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-08T03:46:08,220 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:08,235 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-08T03:46:08,237 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-08T03:46:08,242 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-08T03:46:08,266 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:08,266 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:08,269 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:54406 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:45143:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54406 dst: /127.0.0.1:45143 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:08,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-08T03:46:08,275 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:08,291 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store 2024-12-08T03:46:08,304 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:08,304 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:08,307 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:33164 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:35923:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33164 dst: /127.0.0.1:35923 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:08,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-08T03:46:08,313 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:08,316 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-08T03:46:08,319 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:08,320 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-08T03:46:08,320 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:08,320 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:08,321 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-08T03:46:08,322 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:08,322 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:08,323 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733629568320Disabling compacts and flushes for region at 1733629568320Disabling writes for close at 1733629568321 (+1 ms)Writing region close event to WAL at 1733629568322 (+1 ms)Closed at 1733629568322 2024-12-08T03:46:08,325 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/.initializing 2024-12-08T03:46:08,326 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/WALs/32df4b4bb439,37833,1733629567036 2024-12-08T03:46:08,335 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-08T03:46:08,350 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C37833%2C1733629567036, suffix=, logDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/WALs/32df4b4bb439,37833,1733629567036, archiveDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/oldWALs, maxLogs=10 2024-12-08T03:46:08,384 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/WALs/32df4b4bb439,37833,1733629567036/32df4b4bb439%2C37833%2C1733629567036.1733629568355, exclude list is [], retry=0 2024-12-08T03:46:08,400 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:08,401 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45143,DS-6a00e5d0-2bfe-4ecf-871f-2bff8252b118,DISK] 2024-12-08T03:46:08,401 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38097,DS-58eda793-1940-40cf-a8ce-438598725779,DISK] 2024-12-08T03:46:08,401 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35923,DS-f59ef6c5-2a7a-47c5-8382-e2da01c985c3,DISK] 2024-12-08T03:46:08,404 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-08T03:46:08,441 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/WALs/32df4b4bb439,37833,1733629567036/32df4b4bb439%2C37833%2C1733629567036.1733629568355 2024-12-08T03:46:08,442 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:39763:39763),(127.0.0.1/127.0.0.1:44921:44921),(127.0.0.1/127.0.0.1:38593:38593)] 2024-12-08T03:46:08,442 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-08T03:46:08,443 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:08,445 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,446 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,479 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,499 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-08T03:46:08,502 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:08,504 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:08,505 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,508 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-08T03:46:08,508 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:08,509 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-08T03:46:08,509 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,512 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-08T03:46:08,512 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:08,513 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-08T03:46:08,513 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,515 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-08T03:46:08,515 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:08,516 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-08T03:46:08,517 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,519 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,521 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,525 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,525 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,528 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-08T03:46:08,532 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:08,538 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-08T03:46:08,539 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=59102505, jitterRate=-0.11930404603481293}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-08T03:46:08,545 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733629568456Initializing all the Stores at 1733629568458 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629568458Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629568459 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629568459Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629568459Cleaning up temporary data from old regions at 1733629568525 (+66 ms)Region opened successfully at 1733629568545 (+20 ms) 2024-12-08T03:46:08,546 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-08T03:46:08,576 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5ebd4e06, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32df4b4bb439/172.17.0.2:0 2024-12-08T03:46:08,602 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-08T03:46:08,611 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-08T03:46:08,611 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-08T03:46:08,614 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-08T03:46:08,615 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-08T03:46:08,620 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 4 msec 2024-12-08T03:46:08,620 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-08T03:46:08,645 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-08T03:46:08,654 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-08T03:46:08,690 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-08T03:46:08,693 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-08T03:46:08,695 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-08T03:46:08,706 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-08T03:46:08,708 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-08T03:46:08,711 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-08T03:46:08,716 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-08T03:46:08,717 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-08T03:46:08,724 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-08T03:46:08,743 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-08T03:46:08,749 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-08T03:46:08,758 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:08,758 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:08,758 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:08,758 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:08,758 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,758 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,758 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,758 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,762 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=32df4b4bb439,37833,1733629567036, sessionid=0x10003959ba10000, setting cluster-up flag (Was=false) 2024-12-08T03:46:08,791 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,791 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,791 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,791 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,817 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-08T03:46:08,821 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=32df4b4bb439,37833,1733629567036 2024-12-08T03:46:08,841 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,841 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,841 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,841 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:08,866 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-08T03:46:08,870 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=32df4b4bb439,37833,1733629567036 2024-12-08T03:46:08,878 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-08T03:46:08,943 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-08T03:46:08,951 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-08T03:46:08,958 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-08T03:46:08,960 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(746): ClusterId : 0a3cdf63-875c-4db4-907e-913e10d7eee1 2024-12-08T03:46:08,960 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(746): ClusterId : 0a3cdf63-875c-4db4-907e-913e10d7eee1 2024-12-08T03:46:08,960 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(746): ClusterId : 0a3cdf63-875c-4db4-907e-913e10d7eee1 2024-12-08T03:46:08,962 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-08T03:46:08,962 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-08T03:46:08,962 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-08T03:46:08,965 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 32df4b4bb439,37833,1733629567036 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-08T03:46:08,983 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-08T03:46:08,983 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-08T03:46:08,983 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-08T03:46:08,984 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-08T03:46:08,984 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-08T03:46:08,984 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-08T03:46:08,984 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/32df4b4bb439:0, corePoolSize=5, maxPoolSize=5 2024-12-08T03:46:08,984 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/32df4b4bb439:0, corePoolSize=5, maxPoolSize=5 2024-12-08T03:46:08,985 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/32df4b4bb439:0, corePoolSize=5, maxPoolSize=5 2024-12-08T03:46:08,985 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/32df4b4bb439:0, corePoolSize=5, maxPoolSize=5 2024-12-08T03:46:08,985 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/32df4b4bb439:0, corePoolSize=10, maxPoolSize=10 2024-12-08T03:46:08,985 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:08,985 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/32df4b4bb439:0, corePoolSize=2, maxPoolSize=2 2024-12-08T03:46:08,985 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:08,988 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733629598988 2024-12-08T03:46:08,990 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-08T03:46:08,991 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-08T03:46:08,993 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-08T03:46:08,993 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-08T03:46:08,993 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-08T03:46:08,994 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-08T03:46:08,994 DEBUG [RS:2;32df4b4bb439:36577 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2d819693, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32df4b4bb439/172.17.0.2:0 2024-12-08T03:46:08,994 DEBUG [RS:0;32df4b4bb439:38719 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4c350e78, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32df4b4bb439/172.17.0.2:0 2024-12-08T03:46:08,996 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-08T03:46:08,997 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-08T03:46:08,997 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-08T03:46:08,997 DEBUG [RS:1;32df4b4bb439:40985 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1b8c9223, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32df4b4bb439/172.17.0.2:0 2024-12-08T03:46:08,997 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-08T03:46:08,998 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-08T03:46:09,000 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,000 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-08T03:46:09,007 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,013 DEBUG [RS:2;32df4b4bb439:36577 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;32df4b4bb439:36577 2024-12-08T03:46:09,015 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-08T03:46:09,015 DEBUG [RS:0;32df4b4bb439:38719 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;32df4b4bb439:38719 2024-12-08T03:46:09,016 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-08T03:46:09,016 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-08T03:46:09,016 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-08T03:46:09,016 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-08T03:46:09,016 DEBUG [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-08T03:46:09,016 DEBUG [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-08T03:46:09,016 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-08T03:46:09,017 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-08T03:46:09,019 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(2659): reportForDuty to master=32df4b4bb439,37833,1733629567036 with port=38719, startcode=1733629567727 2024-12-08T03:46:09,019 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(2659): reportForDuty to master=32df4b4bb439,37833,1733629567036 with port=36577, startcode=1733629567903 2024-12-08T03:46:09,021 DEBUG [RS:1;32df4b4bb439:40985 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;32df4b4bb439:40985 2024-12-08T03:46:09,021 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-08T03:46:09,021 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-08T03:46:09,022 DEBUG [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-08T03:46:09,022 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:09,022 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:09,023 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(2659): reportForDuty to master=32df4b4bb439,37833,1733629567036 with port=40985, startcode=1733629567830 2024-12-08T03:46:09,027 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-08T03:46:09,027 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-08T03:46:09,031 DEBUG [RS:2;32df4b4bb439:36577 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-08T03:46:09,031 DEBUG [RS:1;32df4b4bb439:40985 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-08T03:46:09,031 DEBUG [RS:0;32df4b4bb439:38719 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-08T03:46:09,033 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.large.0-1733629569029,5,FailOnTimeoutGroup] 2024-12-08T03:46:09,035 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.small.0-1733629569034,5,FailOnTimeoutGroup] 2024-12-08T03:46:09,035 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,035 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-08T03:46:09,035 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:46710 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:38097:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46710 dst: /127.0.0.1:38097 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:09,036 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,037 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-08T03:46:09,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-08T03:46:09,042 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-08T03:46:09,043 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:09,044 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-08T03:46:09,044 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791 2024-12-08T03:46:09,058 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:09,058 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:09,072 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:46730 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:38097:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46730 dst: /127.0.0.1:38097 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:09,073 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42585, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-08T03:46:09,073 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46447, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-08T03:46:09,074 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52967, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-08T03:46:09,079 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37833 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32df4b4bb439,40985,1733629567830 2024-12-08T03:46:09,081 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-08T03:46:09,082 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37833 {}] master.ServerManager(517): Registering regionserver=32df4b4bb439,40985,1733629567830 2024-12-08T03:46:09,082 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:09,083 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:09,087 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-08T03:46:09,090 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-08T03:46:09,090 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,092 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:09,092 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37833 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32df4b4bb439,36577,1733629567903 2024-12-08T03:46:09,092 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-08T03:46:09,092 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37833 {}] master.ServerManager(517): Registering regionserver=32df4b4bb439,36577,1733629567903 2024-12-08T03:46:09,096 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-08T03:46:09,096 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,096 DEBUG [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791 2024-12-08T03:46:09,096 DEBUG [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42793 2024-12-08T03:46:09,096 DEBUG [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-08T03:46:09,097 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37833 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,097 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37833 {}] master.ServerManager(517): Registering regionserver=32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,097 DEBUG [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791 2024-12-08T03:46:09,098 DEBUG [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42793 2024-12-08T03:46:09,098 DEBUG [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-08T03:46:09,100 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:09,101 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-08T03:46:09,101 DEBUG [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791 2024-12-08T03:46:09,101 DEBUG [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42793 2024-12-08T03:46:09,101 DEBUG [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-08T03:46:09,104 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-08T03:46:09,104 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,105 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:09,105 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-08T03:46:09,108 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-08T03:46:09,108 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,109 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:09,109 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-08T03:46:09,111 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740 2024-12-08T03:46:09,111 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740 2024-12-08T03:46:09,114 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-08T03:46:09,114 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-08T03:46:09,115 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-08T03:46:09,118 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-08T03:46:09,125 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-08T03:46:09,126 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72773644, jitterRate=0.08441179990768433}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-08T03:46:09,128 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733629569083Initializing all the Stores at 1733629569086 (+3 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629569086Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629569087 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629569087Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629569087Cleaning up temporary data from old regions at 1733629569114 (+27 ms)Region opened successfully at 1733629569128 (+14 ms) 2024-12-08T03:46:09,129 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-08T03:46:09,129 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-08T03:46:09,129 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-08T03:46:09,129 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-08T03:46:09,129 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-08T03:46:09,130 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-08T03:46:09,130 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733629569129Disabling compacts and flushes for region at 1733629569129Disabling writes for close at 1733629569129Writing region close event to WAL at 1733629569130 (+1 ms)Closed at 1733629569130 2024-12-08T03:46:09,133 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-08T03:46:09,133 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-08T03:46:09,139 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-08T03:46:09,145 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-08T03:46:09,146 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-08T03:46:09,150 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-08T03:46:09,158 DEBUG [RS:2;32df4b4bb439:36577 {}] zookeeper.ZKUtil(111): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32df4b4bb439,36577,1733629567903 2024-12-08T03:46:09,158 DEBUG [RS:1;32df4b4bb439:40985 {}] zookeeper.ZKUtil(111): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32df4b4bb439,40985,1733629567830 2024-12-08T03:46:09,158 WARN [RS:1;32df4b4bb439:40985 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-08T03:46:09,158 WARN [RS:2;32df4b4bb439:36577 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-08T03:46:09,159 INFO [RS:1;32df4b4bb439:40985 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-08T03:46:09,159 INFO [RS:2;32df4b4bb439:36577 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-08T03:46:09,159 DEBUG [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,40985,1733629567830 2024-12-08T03:46:09,159 DEBUG [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,36577,1733629567903 2024-12-08T03:46:09,160 DEBUG [RS:0;32df4b4bb439:38719 {}] zookeeper.ZKUtil(111): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,160 WARN [RS:0;32df4b4bb439:38719 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-08T03:46:09,160 INFO [RS:0;32df4b4bb439:38719 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-08T03:46:09,160 DEBUG [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,160 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32df4b4bb439,40985,1733629567830] 2024-12-08T03:46:09,160 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32df4b4bb439,36577,1733629567903] 2024-12-08T03:46:09,160 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32df4b4bb439,38719,1733629567727] 2024-12-08T03:46:09,185 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-08T03:46:09,185 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-08T03:46:09,185 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-08T03:46:09,198 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-08T03:46:09,198 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-08T03:46:09,198 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-08T03:46:09,202 INFO [RS:1;32df4b4bb439:40985 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-08T03:46:09,202 INFO [RS:0;32df4b4bb439:38719 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-08T03:46:09,202 INFO [RS:2;32df4b4bb439:36577 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-08T03:46:09,203 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,203 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,203 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,204 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-08T03:46:09,204 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-08T03:46:09,207 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-08T03:46:09,210 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-08T03:46:09,210 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-08T03:46:09,211 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,211 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,211 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-08T03:46:09,212 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,212 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32df4b4bb439:0, corePoolSize=2, maxPoolSize=2 2024-12-08T03:46:09,212 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32df4b4bb439:0, corePoolSize=2, maxPoolSize=2 2024-12-08T03:46:09,212 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32df4b4bb439:0, corePoolSize=2, maxPoolSize=2 2024-12-08T03:46:09,212 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,212 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:09,213 DEBUG [RS:0;32df4b4bb439:38719 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:09,213 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:09,213 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:09,213 DEBUG [RS:1;32df4b4bb439:40985 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:09,213 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:09,213 DEBUG [RS:2;32df4b4bb439:36577 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:09,216 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,216 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,216 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,216 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,216 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,216 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,36577,1733629567903-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-08T03:46:09,216 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,217 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,217 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,217 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,217 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,217 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,38719,1733629567727-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-08T03:46:09,218 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,218 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,218 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,218 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,219 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,219 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,40985,1733629567830-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-08T03:46:09,241 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-08T03:46:09,242 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-08T03:46:09,242 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-08T03:46:09,244 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,38719,1733629567727-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,244 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,36577,1733629567903-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,244 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,40985,1733629567830-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,244 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,244 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,244 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,244 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.Replication(171): 32df4b4bb439,40985,1733629567830 started 2024-12-08T03:46:09,244 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.Replication(171): 32df4b4bb439,36577,1733629567903 started 2024-12-08T03:46:09,244 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.Replication(171): 32df4b4bb439,38719,1733629567727 started 2024-12-08T03:46:09,266 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,266 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,266 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:09,267 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(1482): Serving as 32df4b4bb439,36577,1733629567903, RpcServer on 32df4b4bb439/172.17.0.2:36577, sessionid=0x10003959ba10003 2024-12-08T03:46:09,267 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1482): Serving as 32df4b4bb439,38719,1733629567727, RpcServer on 32df4b4bb439/172.17.0.2:38719, sessionid=0x10003959ba10001 2024-12-08T03:46:09,267 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(1482): Serving as 32df4b4bb439,40985,1733629567830, RpcServer on 32df4b4bb439/172.17.0.2:40985, sessionid=0x10003959ba10002 2024-12-08T03:46:09,267 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-08T03:46:09,267 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-08T03:46:09,267 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-08T03:46:09,268 DEBUG [RS:0;32df4b4bb439:38719 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,268 DEBUG [RS:2;32df4b4bb439:36577 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32df4b4bb439,36577,1733629567903 2024-12-08T03:46:09,268 DEBUG [RS:1;32df4b4bb439:40985 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32df4b4bb439,40985,1733629567830 2024-12-08T03:46:09,268 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,38719,1733629567727' 2024-12-08T03:46:09,268 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,36577,1733629567903' 2024-12-08T03:46:09,268 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,40985,1733629567830' 2024-12-08T03:46:09,268 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-08T03:46:09,268 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-08T03:46:09,268 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-08T03:46:09,269 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-08T03:46:09,269 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-08T03:46:09,269 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-08T03:46:09,270 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-08T03:46:09,270 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-08T03:46:09,270 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-08T03:46:09,270 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-08T03:46:09,270 DEBUG [RS:2;32df4b4bb439:36577 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32df4b4bb439,36577,1733629567903 2024-12-08T03:46:09,270 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-08T03:46:09,270 DEBUG [RS:0;32df4b4bb439:38719 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,270 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,36577,1733629567903' 2024-12-08T03:46:09,270 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-08T03:46:09,270 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-08T03:46:09,270 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,38719,1733629567727' 2024-12-08T03:46:09,270 DEBUG [RS:1;32df4b4bb439:40985 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32df4b4bb439,40985,1733629567830 2024-12-08T03:46:09,270 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-08T03:46:09,270 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,40985,1733629567830' 2024-12-08T03:46:09,270 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-08T03:46:09,271 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-08T03:46:09,271 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-08T03:46:09,271 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-08T03:46:09,271 DEBUG [RS:2;32df4b4bb439:36577 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-08T03:46:09,271 DEBUG [RS:0;32df4b4bb439:38719 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-08T03:46:09,271 DEBUG [RS:1;32df4b4bb439:40985 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-08T03:46:09,271 INFO [RS:0;32df4b4bb439:38719 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-08T03:46:09,271 INFO [RS:2;32df4b4bb439:36577 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-08T03:46:09,271 INFO [RS:1;32df4b4bb439:40985 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-08T03:46:09,271 INFO [RS:0;32df4b4bb439:38719 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-08T03:46:09,271 INFO [RS:2;32df4b4bb439:36577 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-08T03:46:09,271 INFO [RS:1;32df4b4bb439:40985 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-08T03:46:09,300 WARN [32df4b4bb439:37833 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-08T03:46:09,379 INFO [RS:0;32df4b4bb439:38719 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-08T03:46:09,379 INFO [RS:2;32df4b4bb439:36577 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-08T03:46:09,379 INFO [RS:1;32df4b4bb439:40985 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-08T03:46:09,383 INFO [RS:2;32df4b4bb439:36577 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C36577%2C1733629567903, suffix=, logDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,36577,1733629567903, archiveDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/oldWALs, maxLogs=32 2024-12-08T03:46:09,383 INFO [RS:0;32df4b4bb439:38719 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C38719%2C1733629567727, suffix=, logDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,38719,1733629567727, archiveDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/oldWALs, maxLogs=32 2024-12-08T03:46:09,383 INFO [RS:1;32df4b4bb439:40985 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C40985%2C1733629567830, suffix=, logDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,40985,1733629567830, archiveDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/oldWALs, maxLogs=32 2024-12-08T03:46:09,398 DEBUG [RS:2;32df4b4bb439:36577 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,36577,1733629567903/32df4b4bb439%2C36577%2C1733629567903.1733629569386, exclude list is [], retry=0 2024-12-08T03:46:09,400 DEBUG [RS:1;32df4b4bb439:40985 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,40985,1733629567830/32df4b4bb439%2C40985%2C1733629567830.1733629569386, exclude list is [], retry=0 2024-12-08T03:46:09,402 DEBUG [RS:0;32df4b4bb439:38719 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,38719,1733629567727/32df4b4bb439%2C38719%2C1733629567727.1733629569386, exclude list is [], retry=0 2024-12-08T03:46:09,404 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38097,DS-58eda793-1940-40cf-a8ce-438598725779,DISK] 2024-12-08T03:46:09,404 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45143,DS-6a00e5d0-2bfe-4ecf-871f-2bff8252b118,DISK] 2024-12-08T03:46:09,404 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35923,DS-f59ef6c5-2a7a-47c5-8382-e2da01c985c3,DISK] 2024-12-08T03:46:09,405 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45143,DS-6a00e5d0-2bfe-4ecf-871f-2bff8252b118,DISK] 2024-12-08T03:46:09,427 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38097,DS-58eda793-1940-40cf-a8ce-438598725779,DISK] 2024-12-08T03:46:09,427 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35923,DS-f59ef6c5-2a7a-47c5-8382-e2da01c985c3,DISK] 2024-12-08T03:46:09,429 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45143,DS-6a00e5d0-2bfe-4ecf-871f-2bff8252b118,DISK] 2024-12-08T03:46:09,429 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35923,DS-f59ef6c5-2a7a-47c5-8382-e2da01c985c3,DISK] 2024-12-08T03:46:09,430 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38097,DS-58eda793-1940-40cf-a8ce-438598725779,DISK] 2024-12-08T03:46:09,434 INFO [RS:2;32df4b4bb439:36577 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,36577,1733629567903/32df4b4bb439%2C36577%2C1733629567903.1733629569386 2024-12-08T03:46:09,437 INFO [RS:0;32df4b4bb439:38719 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,38719,1733629567727/32df4b4bb439%2C38719%2C1733629567727.1733629569386 2024-12-08T03:46:09,438 DEBUG [RS:2;32df4b4bb439:36577 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:38593:38593),(127.0.0.1/127.0.0.1:44921:44921),(127.0.0.1/127.0.0.1:39763:39763)] 2024-12-08T03:46:09,438 INFO [RS:1;32df4b4bb439:40985 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,40985,1733629567830/32df4b4bb439%2C40985%2C1733629567830.1733629569386 2024-12-08T03:46:09,438 DEBUG [RS:0;32df4b4bb439:38719 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:39763:39763),(127.0.0.1/127.0.0.1:44921:44921),(127.0.0.1/127.0.0.1:38593:38593)] 2024-12-08T03:46:09,439 DEBUG [RS:1;32df4b4bb439:40985 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:38593:38593),(127.0.0.1/127.0.0.1:44921:44921),(127.0.0.1/127.0.0.1:39763:39763)] 2024-12-08T03:46:09,554 DEBUG [32df4b4bb439:37833 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-08T03:46:09,566 DEBUG [32df4b4bb439:37833 {}] balancer.BalancerClusterState(204): Hosts are {32df4b4bb439=0} racks are {/default-rack=0} 2024-12-08T03:46:09,572 DEBUG [32df4b4bb439:37833 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-08T03:46:09,572 DEBUG [32df4b4bb439:37833 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-08T03:46:09,572 DEBUG [32df4b4bb439:37833 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-08T03:46:09,572 DEBUG [32df4b4bb439:37833 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-08T03:46:09,572 DEBUG [32df4b4bb439:37833 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-08T03:46:09,572 DEBUG [32df4b4bb439:37833 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-08T03:46:09,572 INFO [32df4b4bb439:37833 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-08T03:46:09,572 INFO [32df4b4bb439:37833 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-08T03:46:09,572 INFO [32df4b4bb439:37833 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-08T03:46:09,572 DEBUG [32df4b4bb439:37833 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-08T03:46:09,579 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,585 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 32df4b4bb439,38719,1733629567727, state=OPENING 2024-12-08T03:46:09,639 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-08T03:46:09,650 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:09,650 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:09,650 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:09,650 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:09,652 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:09,652 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:09,652 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:09,652 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:09,655 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-08T03:46:09,658 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=32df4b4bb439,38719,1733629567727}] 2024-12-08T03:46:09,836 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-08T03:46:09,839 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49657, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-08T03:46:09,850 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-08T03:46:09,851 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-08T03:46:09,852 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-08T03:46:09,855 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C38719%2C1733629567727.meta, suffix=.meta, logDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,38719,1733629567727, archiveDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/oldWALs, maxLogs=32 2024-12-08T03:46:09,872 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,38719,1733629567727/32df4b4bb439%2C38719%2C1733629567727.meta.1733629569857.meta, exclude list is [], retry=0 2024-12-08T03:46:09,876 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35923,DS-f59ef6c5-2a7a-47c5-8382-e2da01c985c3,DISK] 2024-12-08T03:46:09,876 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:38097,DS-58eda793-1940-40cf-a8ce-438598725779,DISK] 2024-12-08T03:46:09,876 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45143,DS-6a00e5d0-2bfe-4ecf-871f-2bff8252b118,DISK] 2024-12-08T03:46:09,879 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,38719,1733629567727/32df4b4bb439%2C38719%2C1733629567727.meta.1733629569857.meta 2024-12-08T03:46:09,879 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:39763:39763),(127.0.0.1/127.0.0.1:44921:44921),(127.0.0.1/127.0.0.1:38593:38593)] 2024-12-08T03:46:09,879 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-08T03:46:09,881 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-08T03:46:09,883 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-08T03:46:09,888 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-08T03:46:09,891 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-08T03:46:09,892 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:09,892 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-08T03:46:09,892 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-08T03:46:09,895 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-08T03:46:09,896 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-08T03:46:09,896 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,897 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:09,897 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-08T03:46:09,899 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-08T03:46:09,899 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,900 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:09,900 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-08T03:46:09,902 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-08T03:46:09,902 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,903 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:09,903 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-08T03:46:09,905 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-08T03:46:09,905 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:09,906 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:09,906 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-08T03:46:09,907 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740 2024-12-08T03:46:09,910 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740 2024-12-08T03:46:09,912 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-08T03:46:09,912 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-08T03:46:09,913 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-08T03:46:09,915 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-08T03:46:09,916 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63944773, jitterRate=-0.047148630023002625}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-08T03:46:09,917 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-08T03:46:09,918 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733629569892Writing region info on filesystem at 1733629569892Initializing all the Stores at 1733629569894 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629569894Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629569895 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629569895Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629569895Cleaning up temporary data from old regions at 1733629569912 (+17 ms)Running coprocessor post-open hooks at 1733629569917 (+5 ms)Region opened successfully at 1733629569918 (+1 ms) 2024-12-08T03:46:09,924 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733629569828 2024-12-08T03:46:09,934 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-08T03:46:09,935 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-08T03:46:09,936 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,938 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 32df4b4bb439,38719,1733629567727, state=OPEN 2024-12-08T03:46:09,958 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-08T03:46:09,958 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-08T03:46:09,958 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-08T03:46:09,958 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-08T03:46:09,958 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:09,958 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:09,958 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:09,958 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:09,959 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=32df4b4bb439,38719,1733629567727 2024-12-08T03:46:09,966 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-08T03:46:09,966 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=32df4b4bb439,38719,1733629567727 in 302 msec 2024-12-08T03:46:09,973 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-08T03:46:09,973 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 830 msec 2024-12-08T03:46:09,975 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-08T03:46:09,975 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-08T03:46:09,993 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-08T03:46:09,994 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=32df4b4bb439,38719,1733629567727, seqNum=-1] 2024-12-08T03:46:10,033 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-08T03:46:10,035 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58653, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-08T03:46:10,058 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1470 sec 2024-12-08T03:46:10,058 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733629570058, completionTime=-1 2024-12-08T03:46:10,061 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-08T03:46:10,061 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-08T03:46:10,086 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-08T03:46:10,086 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733629630086 2024-12-08T03:46:10,086 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733629690086 2024-12-08T03:46:10,086 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 25 msec 2024-12-08T03:46:10,088 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-08T03:46:10,094 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37833,1733629567036-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:10,094 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37833,1733629567036-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:10,094 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37833,1733629567036-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:10,096 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-32df4b4bb439:37833, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:10,096 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:10,097 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:10,102 DEBUG [master/32df4b4bb439:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-08T03:46:10,121 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.094sec 2024-12-08T03:46:10,122 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-08T03:46:10,123 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-08T03:46:10,123 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-08T03:46:10,124 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-08T03:46:10,124 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-08T03:46:10,125 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37833,1733629567036-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-08T03:46:10,125 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37833,1733629567036-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-08T03:46:10,129 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-08T03:46:10,130 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-08T03:46:10,130 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37833,1733629567036-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:10,173 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@55b943e3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-08T03:46:10,177 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-08T03:46:10,177 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-08T03:46:10,180 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 32df4b4bb439,37833,-1 for getting cluster id 2024-12-08T03:46:10,182 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-08T03:46:10,189 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '0a3cdf63-875c-4db4-907e-913e10d7eee1' 2024-12-08T03:46:10,191 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-08T03:46:10,191 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "0a3cdf63-875c-4db4-907e-913e10d7eee1" 2024-12-08T03:46:10,191 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4eaa85b3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-08T03:46:10,191 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [32df4b4bb439,37833,-1] 2024-12-08T03:46:10,194 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-08T03:46:10,195 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:10,196 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58904, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-08T03:46:10,198 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4d1b8427, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-08T03:46:10,198 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-08T03:46:10,204 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=32df4b4bb439,38719,1733629567727, seqNum=-1] 2024-12-08T03:46:10,204 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-08T03:46:10,206 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58532, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-08T03:46:10,225 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=32df4b4bb439,37833,1733629567036 2024-12-08T03:46:10,228 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-08T03:46:10,233 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 32df4b4bb439,37833,1733629567036 2024-12-08T03:46:10,234 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@4896534d 2024-12-08T03:46:10,235 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-08T03:46:10,237 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58908, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-08T03:46:10,243 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-08T03:46:10,251 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-08T03:46:10,254 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-08T03:46:10,256 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-08T03:46:10,256 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:10,258 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-08T03:46:10,261 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-08T03:46:10,265 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:10,265 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:10,269 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:46772 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:38097:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46772 dst: /127.0.0.1:38097 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:10,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-08T03:46:10,275 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:10,278 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => ffb51ea0beb02db11b45cf029af18a18, NAME => 'TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791 2024-12-08T03:46:10,283 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:10,283 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:10,286 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:54466 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:45143:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54466 dst: /127.0.0.1:45143 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:10,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-08T03:46:10,293 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:10,294 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:10,294 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing ffb51ea0beb02db11b45cf029af18a18, disabling compactions & flushes 2024-12-08T03:46:10,294 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:10,294 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:10,294 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. after waiting 0 ms 2024-12-08T03:46:10,294 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:10,294 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:10,294 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for ffb51ea0beb02db11b45cf029af18a18: Waiting for close lock at 1733629570294Disabling compacts and flushes for region at 1733629570294Disabling writes for close at 1733629570294Writing region close event to WAL at 1733629570294Closed at 1733629570294 2024-12-08T03:46:10,296 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-08T03:46:10,301 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733629570297"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733629570297"}]},"ts":"1733629570297"} 2024-12-08T03:46:10,305 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-08T03:46:10,307 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-08T03:46:10,310 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733629570307"}]},"ts":"1733629570307"} 2024-12-08T03:46:10,315 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-08T03:46:10,315 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {32df4b4bb439=0} racks are {/default-rack=0} 2024-12-08T03:46:10,316 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-08T03:46:10,317 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-08T03:46:10,317 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-08T03:46:10,317 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-08T03:46:10,317 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-08T03:46:10,317 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-08T03:46:10,317 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-08T03:46:10,317 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-08T03:46:10,317 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-08T03:46:10,317 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-08T03:46:10,318 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ffb51ea0beb02db11b45cf029af18a18, ASSIGN}] 2024-12-08T03:46:10,321 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ffb51ea0beb02db11b45cf029af18a18, ASSIGN 2024-12-08T03:46:10,323 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ffb51ea0beb02db11b45cf029af18a18, ASSIGN; state=OFFLINE, location=32df4b4bb439,38719,1733629567727; forceNewPlan=false, retain=false 2024-12-08T03:46:10,372 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-08T03:46:10,477 INFO [32df4b4bb439:37833 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-08T03:46:10,478 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=ffb51ea0beb02db11b45cf029af18a18, regionState=OPENING, regionLocation=32df4b4bb439,38719,1733629567727 2024-12-08T03:46:10,484 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ffb51ea0beb02db11b45cf029af18a18, ASSIGN because future has completed 2024-12-08T03:46:10,485 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure ffb51ea0beb02db11b45cf029af18a18, server=32df4b4bb439,38719,1733629567727}] 2024-12-08T03:46:10,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-08T03:46:10,646 INFO [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:10,646 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => ffb51ea0beb02db11b45cf029af18a18, NAME => 'TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18.', STARTKEY => '', ENDKEY => ''} 2024-12-08T03:46:10,647 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,647 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:10,647 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,647 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,650 INFO [StoreOpener-ffb51ea0beb02db11b45cf029af18a18-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,652 INFO [StoreOpener-ffb51ea0beb02db11b45cf029af18a18-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region ffb51ea0beb02db11b45cf029af18a18 columnFamilyName cf 2024-12-08T03:46:10,652 DEBUG [StoreOpener-ffb51ea0beb02db11b45cf029af18a18-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:10,653 INFO [StoreOpener-ffb51ea0beb02db11b45cf029af18a18-1 {}] regionserver.HStore(327): Store=ffb51ea0beb02db11b45cf029af18a18/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-08T03:46:10,653 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,655 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,655 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,656 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,656 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,658 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,663 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-08T03:46:10,664 INFO [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened ffb51ea0beb02db11b45cf029af18a18; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72731687, jitterRate=0.08378659188747406}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-08T03:46:10,664 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:10,665 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for ffb51ea0beb02db11b45cf029af18a18: Running coprocessor pre-open hook at 1733629570647Writing region info on filesystem at 1733629570647Initializing all the Stores at 1733629570649 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629570649Cleaning up temporary data from old regions at 1733629570656 (+7 ms)Running coprocessor post-open hooks at 1733629570664 (+8 ms)Region opened successfully at 1733629570665 (+1 ms) 2024-12-08T03:46:10,666 INFO [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18., pid=6, masterSystemTime=1733629570639 2024-12-08T03:46:10,670 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:10,670 INFO [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:10,671 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=ffb51ea0beb02db11b45cf029af18a18, regionState=OPEN, openSeqNum=2, regionLocation=32df4b4bb439,38719,1733629567727 2024-12-08T03:46:10,675 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure ffb51ea0beb02db11b45cf029af18a18, server=32df4b4bb439,38719,1733629567727 because future has completed 2024-12-08T03:46:10,682 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-08T03:46:10,684 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure ffb51ea0beb02db11b45cf029af18a18, server=32df4b4bb439,38719,1733629567727 in 192 msec 2024-12-08T03:46:10,687 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-08T03:46:10,687 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ffb51ea0beb02db11b45cf029af18a18, ASSIGN in 364 msec 2024-12-08T03:46:10,689 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-08T03:46:10,689 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733629570689"}]},"ts":"1733629570689"} 2024-12-08T03:46:10,692 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-08T03:46:10,693 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-08T03:46:10,697 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 448 msec 2024-12-08T03:46:10,892 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-08T03:46:10,894 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-08T03:46:10,894 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-08T03:46:10,895 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-08T03:46:10,902 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-08T03:46:10,902 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-08T03:46:10,903 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-08T03:46:10,913 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18., hostname=32df4b4bb439,38719,1733629567727, seqNum=2] 2024-12-08T03:46:10,923 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-08T03:46:10,927 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-08T03:46:10,929 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-08T03:46:10,929 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-08T03:46:10,931 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-08T03:46:10,932 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-08T03:46:11,032 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-08T03:46:11,101 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=38719 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-08T03:46:11,101 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:11,105 INFO [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing ffb51ea0beb02db11b45cf029af18a18 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-08T03:46:11,152 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18/.tmp/cf/1b6e504523144f349c3b171901d68f19 is 36, key is row/cf:cq/1733629570916/Put/seqid=0 2024-12-08T03:46:11,158 WARN [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,158 WARN [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,161 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_436998923_22 at /127.0.0.1:53300 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:38097:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53300 dst: /127.0.0.1:38097 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:11,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-08T03:46:11,166 WARN [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:11,167 INFO [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18/.tmp/cf/1b6e504523144f349c3b171901d68f19 2024-12-08T03:46:11,206 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18/.tmp/cf/1b6e504523144f349c3b171901d68f19 as hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18/cf/1b6e504523144f349c3b171901d68f19 2024-12-08T03:46:11,217 INFO [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18/cf/1b6e504523144f349c3b171901d68f19, entries=1, sequenceid=5, filesize=4.7 K 2024-12-08T03:46:11,226 INFO [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for ffb51ea0beb02db11b45cf029af18a18 in 118ms, sequenceid=5, compaction requested=false 2024-12-08T03:46:11,227 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-08T03:46:11,229 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for ffb51ea0beb02db11b45cf029af18a18: 2024-12-08T03:46:11,229 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:11,230 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-08T03:46:11,232 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-08T03:46:11,237 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-08T03:46:11,237 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 302 msec 2024-12-08T03:46:11,240 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 314 msec 2024-12-08T03:46:11,242 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37833 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-08T03:46:11,242 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-08T03:46:11,253 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-08T03:46:11,254 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-08T03:46:11,254 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:11,257 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,258 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,258 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-08T03:46:11,258 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-08T03:46:11,258 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=201166883, stopped=false 2024-12-08T03:46:11,258 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=32df4b4bb439,37833,1733629567036 2024-12-08T03:46:11,316 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:11,316 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:11,316 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:11,316 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:11,316 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:11,316 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:11,316 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:11,316 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:11,317 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-08T03:46:11,318 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:11,318 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:11,318 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-08T03:46:11,318 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:11,318 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:11,319 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:11,319 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,319 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32df4b4bb439,38719,1733629567727' ***** 2024-12-08T03:46:11,320 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-08T03:46:11,320 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32df4b4bb439,40985,1733629567830' ***** 2024-12-08T03:46:11,320 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(2210): STOPPED: Exiting; cluster shutdown set and not carrying any regions 2024-12-08T03:46:11,320 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32df4b4bb439,40985,1733629567830' ***** 2024-12-08T03:46:11,320 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32df4b4bb439,36577,1733629567903' ***** 2024-12-08T03:46:11,320 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-08T03:46:11,320 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(2210): STOPPED: Exiting; cluster shutdown set and not carrying any regions 2024-12-08T03:46:11,321 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-08T03:46:11,322 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-08T03:46:11,322 INFO [RS:0;32df4b4bb439:38719 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-08T03:46:11,322 INFO [RS:0;32df4b4bb439:38719 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-08T03:46:11,323 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(3091): Received CLOSE for ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:11,323 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-08T03:46:11,323 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(959): stopping server 32df4b4bb439,38719,1733629567727 2024-12-08T03:46:11,324 INFO [RS:2;32df4b4bb439:36577 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-08T03:46:11,324 INFO [RS:0;32df4b4bb439:38719 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-08T03:46:11,324 INFO [RS:2;32df4b4bb439:36577 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-08T03:46:11,324 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(959): stopping server 32df4b4bb439,36577,1733629567903 2024-12-08T03:46:11,324 INFO [RS:0;32df4b4bb439:38719 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;32df4b4bb439:38719. 2024-12-08T03:46:11,324 INFO [RS:2;32df4b4bb439:36577 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-08T03:46:11,324 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-08T03:46:11,324 DEBUG [RS:0;32df4b4bb439:38719 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:11,324 INFO [RS:2;32df4b4bb439:36577 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;32df4b4bb439:36577. 2024-12-08T03:46:11,324 DEBUG [RS:0;32df4b4bb439:38719 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,324 INFO [RS:1;32df4b4bb439:40985 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-08T03:46:11,324 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-08T03:46:11,324 DEBUG [RS:2;32df4b4bb439:36577 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:11,324 INFO [RS:1;32df4b4bb439:40985 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-08T03:46:11,324 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-08T03:46:11,324 DEBUG [RS:2;32df4b4bb439:36577 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,324 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(959): stopping server 32df4b4bb439,40985,1733629567830 2024-12-08T03:46:11,324 INFO [RS:1;32df4b4bb439:40985 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-08T03:46:11,324 INFO [RS:1;32df4b4bb439:40985 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;32df4b4bb439:40985. 2024-12-08T03:46:11,325 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(976): stopping server 32df4b4bb439,36577,1733629567903; all regions closed. 2024-12-08T03:46:11,325 DEBUG [RS:1;32df4b4bb439:40985 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:11,325 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-08T03:46:11,325 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing ffb51ea0beb02db11b45cf029af18a18, disabling compactions & flushes 2024-12-08T03:46:11,325 DEBUG [RS:1;32df4b4bb439:40985 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,325 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-08T03:46:11,325 INFO [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:11,325 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-08T03:46:11,325 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(976): stopping server 32df4b4bb439,40985,1733629567830; all regions closed. 2024-12-08T03:46:11,325 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:11,325 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-08T03:46:11,325 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. after waiting 0 ms 2024-12-08T03:46:11,325 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:11,326 INFO [regionserver/32df4b4bb439:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:11,326 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-08T03:46:11,326 DEBUG [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740, ffb51ea0beb02db11b45cf029af18a18=TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18.} 2024-12-08T03:46:11,326 DEBUG [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, ffb51ea0beb02db11b45cf029af18a18 2024-12-08T03:46:11,326 INFO [regionserver/32df4b4bb439:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:11,326 INFO [regionserver/32df4b4bb439:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:11,327 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-08T03:46:11,327 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-08T03:46:11,327 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-08T03:46:11,327 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-08T03:46:11,327 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-08T03:46:11,328 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-08T03:46:11,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_1073741826_1016 (size=93) 2024-12-08T03:46:11,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_1073741827_1017 (size=93) 2024-12-08T03:46:11,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_1073741826_1016 (size=93) 2024-12-08T03:46:11,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_1073741827_1017 (size=93) 2024-12-08T03:46:11,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_1073741826_1016 (size=93) 2024-12-08T03:46:11,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_1073741827_1017 (size=93) 2024-12-08T03:46:11,340 DEBUG [RS:1;32df4b4bb439:40985 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/oldWALs 2024-12-08T03:46:11,340 DEBUG [RS:2;32df4b4bb439:36577 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/oldWALs 2024-12-08T03:46:11,340 INFO [RS:1;32df4b4bb439:40985 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 32df4b4bb439%2C40985%2C1733629567830:(num 1733629569386) 2024-12-08T03:46:11,340 INFO [RS:2;32df4b4bb439:36577 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 32df4b4bb439%2C36577%2C1733629567903:(num 1733629569386) 2024-12-08T03:46:11,340 DEBUG [RS:1;32df4b4bb439:40985 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,340 DEBUG [RS:2;32df4b4bb439:36577 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,340 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:11,340 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:11,340 INFO [RS:1;32df4b4bb439:40985 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-08T03:46:11,340 INFO [RS:2;32df4b4bb439:36577 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-08T03:46:11,341 INFO [RS:1;32df4b4bb439:40985 {}] hbase.ChoreService(370): Chore service for: regionserver/32df4b4bb439:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-08T03:46:11,341 INFO [RS:2;32df4b4bb439:36577 {}] hbase.ChoreService(370): Chore service for: regionserver/32df4b4bb439:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-08T03:46:11,341 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-08T03:46:11,341 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-08T03:46:11,341 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-08T03:46:11,341 INFO [regionserver/32df4b4bb439:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-08T03:46:11,341 INFO [regionserver/32df4b4bb439:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-08T03:46:11,341 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-08T03:46:11,341 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-08T03:46:11,341 INFO [RS:1;32df4b4bb439:40985 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-08T03:46:11,341 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-08T03:46:11,341 INFO [RS:2;32df4b4bb439:36577 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-08T03:46:11,341 INFO [RS:1;32df4b4bb439:40985 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:40985 2024-12-08T03:46:11,341 INFO [RS:2;32df4b4bb439:36577 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36577 2024-12-08T03:46:11,348 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/default/TestHBaseWalOnEC/ffb51ea0beb02db11b45cf029af18a18/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-08T03:46:11,349 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-08T03:46:11,349 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32df4b4bb439,36577,1733629567903 2024-12-08T03:46:11,349 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32df4b4bb439,40985,1733629567830 2024-12-08T03:46:11,349 INFO [RS:2;32df4b4bb439:36577 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-08T03:46:11,349 INFO [RS:1;32df4b4bb439:40985 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-08T03:46:11,351 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32df4b4bb439,40985,1733629567830] 2024-12-08T03:46:11,351 INFO [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:11,351 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for ffb51ea0beb02db11b45cf029af18a18: Waiting for close lock at 1733629571324Running coprocessor pre-close hooks at 1733629571324Disabling compacts and flushes for region at 1733629571324Disabling writes for close at 1733629571325 (+1 ms)Writing region close event to WAL at 1733629571326 (+1 ms)Running coprocessor post-close hooks at 1733629571350 (+24 ms)Closed at 1733629571351 (+1 ms) 2024-12-08T03:46:11,352 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18. 2024-12-08T03:46:11,366 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32df4b4bb439,40985,1733629567830 already deleted, retry=false 2024-12-08T03:46:11,366 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32df4b4bb439,40985,1733629567830 expired; onlineServers=2 2024-12-08T03:46:11,366 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32df4b4bb439,36577,1733629567903] 2024-12-08T03:46:11,372 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/info/7c17632d570a4573a1140049929cd441 is 153, key is TestHBaseWalOnEC,,1733629570239.ffb51ea0beb02db11b45cf029af18a18./info:regioninfo/1733629570670/Put/seqid=0 2024-12-08T03:46:11,374 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32df4b4bb439,36577,1733629567903 already deleted, retry=false 2024-12-08T03:46:11,374 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32df4b4bb439,36577,1733629567903 expired; onlineServers=1 2024-12-08T03:46:11,375 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,375 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,380 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_436998923_22 at /127.0.0.1:53318 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:38097:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53318 dst: /127.0.0.1:38097 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:11,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-08T03:46:11,385 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:11,386 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/info/7c17632d570a4573a1140049929cd441 2024-12-08T03:46:11,412 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/ns/162dd356fdf64010baa9d34495ccf68e is 43, key is default/ns:d/1733629570039/Put/seqid=0 2024-12-08T03:46:11,415 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,415 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,418 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_436998923_22 at /127.0.0.1:33500 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:35923:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33500 dst: /127.0.0.1:35923 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:11,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-08T03:46:11,423 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:11,423 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/ns/162dd356fdf64010baa9d34495ccf68e 2024-12-08T03:46:11,447 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/table/5874150750f746e59b606f175bda0c12 is 52, key is TestHBaseWalOnEC/table:state/1733629570689/Put/seqid=0 2024-12-08T03:46:11,448 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,449 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,451 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_436998923_22 at /127.0.0.1:33508 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:35923:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33508 dst: /127.0.0.1:35923 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:11,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-08T03:46:11,456 WARN [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:11,457 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/table/5874150750f746e59b606f175bda0c12 2024-12-08T03:46:11,458 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:11,458 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40985-0x10003959ba10002, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:11,458 INFO [RS:1;32df4b4bb439:40985 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-08T03:46:11,458 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:11,458 INFO [RS:2;32df4b4bb439:36577 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-08T03:46:11,459 DEBUG [pool-77-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36577-0x10003959ba10003, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:11,459 INFO [RS:1;32df4b4bb439:40985 {}] regionserver.HRegionServer(1031): Exiting; stopping=32df4b4bb439,40985,1733629567830; zookeeper connection closed. 2024-12-08T03:46:11,459 INFO [RS:2;32df4b4bb439:36577 {}] regionserver.HRegionServer(1031): Exiting; stopping=32df4b4bb439,36577,1733629567903; zookeeper connection closed. 2024-12-08T03:46:11,459 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@1f8dcd6f {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@1f8dcd6f 2024-12-08T03:46:11,459 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@15721a2e {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@15721a2e 2024-12-08T03:46:11,467 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/info/7c17632d570a4573a1140049929cd441 as hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/info/7c17632d570a4573a1140049929cd441 2024-12-08T03:46:11,476 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/info/7c17632d570a4573a1140049929cd441, entries=10, sequenceid=11, filesize=6.5 K 2024-12-08T03:46:11,477 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/ns/162dd356fdf64010baa9d34495ccf68e as hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/ns/162dd356fdf64010baa9d34495ccf68e 2024-12-08T03:46:11,486 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/ns/162dd356fdf64010baa9d34495ccf68e, entries=2, sequenceid=11, filesize=5.0 K 2024-12-08T03:46:11,488 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/.tmp/table/5874150750f746e59b606f175bda0c12 as hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/table/5874150750f746e59b606f175bda0c12 2024-12-08T03:46:11,497 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/table/5874150750f746e59b606f175bda0c12, entries=2, sequenceid=11, filesize=5.1 K 2024-12-08T03:46:11,499 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 172ms, sequenceid=11, compaction requested=false 2024-12-08T03:46:11,499 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-08T03:46:11,510 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-08T03:46:11,511 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-08T03:46:11,512 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-08T03:46:11,512 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733629571327Running coprocessor pre-close hooks at 1733629571327Disabling compacts and flushes for region at 1733629571327Disabling writes for close at 1733629571327Obtaining lock to block concurrent updates at 1733629571328 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1733629571328Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733629571329 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733629571330 (+1 ms)Flushing 1588230740/info: creating writer at 1733629571330Flushing 1588230740/info: appending metadata at 1733629571368 (+38 ms)Flushing 1588230740/info: closing flushed file at 1733629571368Flushing 1588230740/ns: creating writer at 1733629571396 (+28 ms)Flushing 1588230740/ns: appending metadata at 1733629571411 (+15 ms)Flushing 1588230740/ns: closing flushed file at 1733629571411Flushing 1588230740/table: creating writer at 1733629571431 (+20 ms)Flushing 1588230740/table: appending metadata at 1733629571446 (+15 ms)Flushing 1588230740/table: closing flushed file at 1733629571446Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@31f36fea: reopening flushed file at 1733629571465 (+19 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@77d6adcd: reopening flushed file at 1733629571476 (+11 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@14bfc16d: reopening flushed file at 1733629571486 (+10 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 172ms, sequenceid=11, compaction requested=false at 1733629571499 (+13 ms)Writing region close event to WAL at 1733629571500 (+1 ms)Running coprocessor post-close hooks at 1733629571511 (+11 ms)Closed at 1733629571512 (+1 ms) 2024-12-08T03:46:11,512 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-08T03:46:11,527 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(976): stopping server 32df4b4bb439,38719,1733629567727; all regions closed. 2024-12-08T03:46:11,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_1073741829_1019 (size=2751) 2024-12-08T03:46:11,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_1073741829_1019 (size=2751) 2024-12-08T03:46:11,530 WARN [Close-WAL-Writer-0 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(650): complete file /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/WALs/32df4b4bb439,38719,1733629567727/32df4b4bb439%2C38719%2C1733629567727.meta.1733629569857.meta not finished, retry = 0 2024-12-08T03:46:11,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_1073741829_1019 (size=2751) 2024-12-08T03:46:11,638 DEBUG [RS:0;32df4b4bb439:38719 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/oldWALs 2024-12-08T03:46:11,639 INFO [RS:0;32df4b4bb439:38719 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 32df4b4bb439%2C38719%2C1733629567727.meta:.meta(num 1733629569857) 2024-12-08T03:46:11,644 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_1073741828_1018 (size=1298) 2024-12-08T03:46:11,644 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_1073741828_1018 (size=1298) 2024-12-08T03:46:11,644 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_1073741828_1018 (size=1298) 2024-12-08T03:46:11,648 DEBUG [RS:0;32df4b4bb439:38719 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/oldWALs 2024-12-08T03:46:11,648 INFO [RS:0;32df4b4bb439:38719 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 32df4b4bb439%2C38719%2C1733629567727:(num 1733629569386) 2024-12-08T03:46:11,648 DEBUG [RS:0;32df4b4bb439:38719 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:11,648 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:11,648 INFO [RS:0;32df4b4bb439:38719 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-08T03:46:11,648 INFO [RS:0;32df4b4bb439:38719 {}] hbase.ChoreService(370): Chore service for: regionserver/32df4b4bb439:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-08T03:46:11,649 INFO [RS:0;32df4b4bb439:38719 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-08T03:46:11,649 INFO [regionserver/32df4b4bb439:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-08T03:46:11,649 INFO [RS:0;32df4b4bb439:38719 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:38719 2024-12-08T03:46:11,717 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32df4b4bb439,38719,1733629567727 2024-12-08T03:46:11,717 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-08T03:46:11,717 INFO [RS:0;32df4b4bb439:38719 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-08T03:46:11,725 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32df4b4bb439,38719,1733629567727] 2024-12-08T03:46:11,733 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32df4b4bb439,38719,1733629567727 already deleted, retry=false 2024-12-08T03:46:11,733 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32df4b4bb439,38719,1733629567727 expired; onlineServers=0 2024-12-08T03:46:11,733 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '32df4b4bb439,37833,1733629567036' ***** 2024-12-08T03:46:11,734 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-08T03:46:11,734 INFO [M:0;32df4b4bb439:37833 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-08T03:46:11,734 INFO [M:0;32df4b4bb439:37833 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-08T03:46:11,734 DEBUG [M:0;32df4b4bb439:37833 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-08T03:46:11,735 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-08T03:46:11,735 DEBUG [M:0;32df4b4bb439:37833 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-08T03:46:11,735 DEBUG [master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.small.0-1733629569034 {}] cleaner.HFileCleaner(306): Exit Thread[master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.small.0-1733629569034,5,FailOnTimeoutGroup] 2024-12-08T03:46:11,735 DEBUG [master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.large.0-1733629569029 {}] cleaner.HFileCleaner(306): Exit Thread[master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.large.0-1733629569029,5,FailOnTimeoutGroup] 2024-12-08T03:46:11,735 INFO [M:0;32df4b4bb439:37833 {}] hbase.ChoreService(370): Chore service for: master/32df4b4bb439:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-08T03:46:11,735 INFO [M:0;32df4b4bb439:37833 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-08T03:46:11,735 DEBUG [M:0;32df4b4bb439:37833 {}] master.HMaster(1795): Stopping service threads 2024-12-08T03:46:11,735 INFO [M:0;32df4b4bb439:37833 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-08T03:46:11,735 INFO [M:0;32df4b4bb439:37833 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-08T03:46:11,736 INFO [M:0;32df4b4bb439:37833 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-08T03:46:11,736 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-08T03:46:11,741 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-08T03:46:11,741 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:11,741 DEBUG [M:0;32df4b4bb439:37833 {}] zookeeper.ZKUtil(347): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-08T03:46:11,741 WARN [M:0;32df4b4bb439:37833 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-08T03:46:11,742 INFO [M:0;32df4b4bb439:37833 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/.lastflushedseqids 2024-12-08T03:46:11,752 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,752 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,755 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:33528 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:35923:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33528 dst: /127.0.0.1:35923 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:11,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-08T03:46:11,759 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:11,759 INFO [M:0;32df4b4bb439:37833 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-08T03:46:11,759 INFO [M:0;32df4b4bb439:37833 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-08T03:46:11,759 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-08T03:46:11,759 INFO [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:11,759 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:11,759 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-08T03:46:11,759 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:11,760 INFO [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.83 KB heapSize=34.12 KB 2024-12-08T03:46:11,778 DEBUG [M:0;32df4b4bb439:37833 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/04e3bd544a2e49d2a0f0bc64fc78d9d4 is 82, key is hbase:meta,,1/info:regioninfo/1733629569936/Put/seqid=0 2024-12-08T03:46:11,780 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,781 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,783 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:33552 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:35923:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33552 dst: /127.0.0.1:35923 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:11,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-08T03:46:11,787 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:11,787 INFO [M:0;32df4b4bb439:37833 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/04e3bd544a2e49d2a0f0bc64fc78d9d4 2024-12-08T03:46:11,809 DEBUG [M:0;32df4b4bb439:37833 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/97e82e77e5754b6495ade4b8cfc0558e is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733629570696/Put/seqid=0 2024-12-08T03:46:11,811 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,811 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,813 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:53330 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:38097:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53330 dst: /127.0.0.1:38097 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:11,817 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775552_1037 (size=6439) 2024-12-08T03:46:11,818 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:11,818 INFO [M:0;32df4b4bb439:37833 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/97e82e77e5754b6495ade4b8cfc0558e 2024-12-08T03:46:11,825 INFO [RS:0;32df4b4bb439:38719 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-08T03:46:11,825 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:11,825 INFO [RS:0;32df4b4bb439:38719 {}] regionserver.HRegionServer(1031): Exiting; stopping=32df4b4bb439,38719,1733629567727; zookeeper connection closed. 2024-12-08T03:46:11,825 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38719-0x10003959ba10001, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:11,825 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@62ffe58e {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@62ffe58e 2024-12-08T03:46:11,826 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-08T03:46:11,840 DEBUG [M:0;32df4b4bb439:37833 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/bbac9a8c289b4b9d9f9dc420d54d1509 is 69, key is 32df4b4bb439,36577,1733629567903/rs:state/1733629569093/Put/seqid=0 2024-12-08T03:46:11,842 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,842 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-08T03:46:11,844 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1237983690_22 at /127.0.0.1:36152 [Receiving block BP-1980098496-172.17.0.2-1733629563223:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:45143:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36152 dst: /127.0.0.1:45143 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-08T03:46:11,848 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-08T03:46:11,848 WARN [M:0;32df4b4bb439:37833 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-08T03:46:11,848 INFO [M:0;32df4b4bb439:37833 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/bbac9a8c289b4b9d9f9dc420d54d1509 2024-12-08T03:46:11,856 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/04e3bd544a2e49d2a0f0bc64fc78d9d4 as hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/04e3bd544a2e49d2a0f0bc64fc78d9d4 2024-12-08T03:46:11,863 INFO [M:0;32df4b4bb439:37833 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/04e3bd544a2e49d2a0f0bc64fc78d9d4, entries=8, sequenceid=72, filesize=5.5 K 2024-12-08T03:46:11,865 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/97e82e77e5754b6495ade4b8cfc0558e as hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/97e82e77e5754b6495ade4b8cfc0558e 2024-12-08T03:46:11,872 INFO [M:0;32df4b4bb439:37833 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/97e82e77e5754b6495ade4b8cfc0558e, entries=8, sequenceid=72, filesize=6.3 K 2024-12-08T03:46:11,874 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/bbac9a8c289b4b9d9f9dc420d54d1509 as hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/bbac9a8c289b4b9d9f9dc420d54d1509 2024-12-08T03:46:11,880 INFO [M:0;32df4b4bb439:37833 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/bbac9a8c289b4b9d9f9dc420d54d1509, entries=3, sequenceid=72, filesize=5.2 K 2024-12-08T03:46:11,882 INFO [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 121ms, sequenceid=72, compaction requested=false 2024-12-08T03:46:11,883 INFO [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:11,883 DEBUG [M:0;32df4b4bb439:37833 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733629571759Disabling compacts and flushes for region at 1733629571759Disabling writes for close at 1733629571759Obtaining lock to block concurrent updates at 1733629571760 (+1 ms)Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733629571760Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27471, getHeapSize=34872, getOffHeapSize=0, getCellsCount=85 at 1733629571760Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733629571761 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733629571761Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733629571778 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733629571778Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733629571794 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733629571809 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733629571809Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733629571824 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733629571839 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733629571839Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4d2095c6: reopening flushed file at 1733629571855 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@24d100e4: reopening flushed file at 1733629571863 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@74c04b1: reopening flushed file at 1733629571872 (+9 ms)Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 121ms, sequenceid=72, compaction requested=false at 1733629571882 (+10 ms)Writing region close event to WAL at 1733629571883 (+1 ms)Closed at 1733629571883 2024-12-08T03:46:11,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_1073741825_1011 (size=32674) 2024-12-08T03:46:11,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_1073741825_1011 (size=32674) 2024-12-08T03:46:11,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_1073741825_1011 (size=32674) 2024-12-08T03:46:11,887 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-08T03:46:11,887 INFO [M:0;32df4b4bb439:37833 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-08T03:46:11,887 INFO [M:0;32df4b4bb439:37833 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:37833 2024-12-08T03:46:11,888 INFO [M:0;32df4b4bb439:37833 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-08T03:46:11,985 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_-9223372036854775725_1010 (size=34) 2024-12-08T03:46:11,985 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775724_1010 (size=34) 2024-12-08T03:46:11,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775773_1004 (size=42) 2024-12-08T03:46:11,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_-9223372036854775772_1004 (size=42) 2024-12-08T03:46:12,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38097 is added to blk_-9223372036854775740_1008 (size=1189) 2024-12-08T03:46:12,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775741_1008 (size=1189) 2024-12-08T03:46:12,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45143 is added to blk_-9223372036854775756_1006 (size=196) 2024-12-08T03:46:12,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35923 is added to blk_-9223372036854775757_1006 (size=196) 2024-12-08T03:46:12,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:12,025 INFO [M:0;32df4b4bb439:37833 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-08T03:46:12,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37833-0x10003959ba10000, quorum=127.0.0.1:56008, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:12,056 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@653e6301{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:12,060 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@404caff2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-08T03:46:12,060 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-08T03:46:12,060 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@343b36c2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-08T03:46:12,061 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@35e2f174{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir/,STOPPED} 2024-12-08T03:46:12,064 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-08T03:46:12,064 WARN [BP-1980098496-172.17.0.2-1733629563223 heartbeating to localhost/127.0.0.1:42793 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-08T03:46:12,064 WARN [BP-1980098496-172.17.0.2-1733629563223 heartbeating to localhost/127.0.0.1:42793 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1980098496-172.17.0.2-1733629563223 (Datanode Uuid 8930704c-b215-4d25-9b2f-8d95d2ea3622) service to localhost/127.0.0.1:42793 2024-12-08T03:46:12,064 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-08T03:46:12,065 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data5/current/BP-1980098496-172.17.0.2-1733629563223 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:12,066 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data6/current/BP-1980098496-172.17.0.2-1733629563223 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:12,066 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-08T03:46:12,069 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@513cab2c{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:12,070 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@29a123ec{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-08T03:46:12,070 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-08T03:46:12,070 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6af5a446{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-08T03:46:12,070 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@444b27d4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir/,STOPPED} 2024-12-08T03:46:12,071 WARN [BP-1980098496-172.17.0.2-1733629563223 heartbeating to localhost/127.0.0.1:42793 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-08T03:46:12,071 WARN [BP-1980098496-172.17.0.2-1733629563223 heartbeating to localhost/127.0.0.1:42793 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1980098496-172.17.0.2-1733629563223 (Datanode Uuid 68cb5d41-74fa-475e-9d87-aa1ee7d5825d) service to localhost/127.0.0.1:42793 2024-12-08T03:46:12,072 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data3/current/BP-1980098496-172.17.0.2-1733629563223 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:12,072 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data4/current/BP-1980098496-172.17.0.2-1733629563223 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:12,072 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-08T03:46:12,072 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-08T03:46:12,072 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-08T03:46:12,074 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@65462677{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:12,074 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@383014b{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-08T03:46:12,074 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-08T03:46:12,074 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4dc262e0{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-08T03:46:12,074 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@431e53b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir/,STOPPED} 2024-12-08T03:46:12,075 WARN [BP-1980098496-172.17.0.2-1733629563223 heartbeating to localhost/127.0.0.1:42793 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-08T03:46:12,075 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-08T03:46:12,075 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-08T03:46:12,075 WARN [BP-1980098496-172.17.0.2-1733629563223 heartbeating to localhost/127.0.0.1:42793 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1980098496-172.17.0.2-1733629563223 (Datanode Uuid f3d0217e-5dbb-44f3-96a2-2e3d2dcf7210) service to localhost/127.0.0.1:42793 2024-12-08T03:46:12,076 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data1/current/BP-1980098496-172.17.0.2-1733629563223 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:12,076 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/cluster_3dfa8485-a6cd-8556-a194-35285e4c3797/data/data2/current/BP-1980098496-172.17.0.2-1733629563223 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:12,076 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-08T03:46:12,084 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@58dbf239{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-08T03:46:12,085 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@13e2962d{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-08T03:46:12,085 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-08T03:46:12,085 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4f93dd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-08T03:46:12,085 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4395d44b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir/,STOPPED} 2024-12-08T03:46:12,092 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-08T03:46:12,118 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-08T03:46:12,124 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=90 (was 160), OpenFileDescriptor=439 (was 393) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=141 (was 136) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=17505 (was 17806) 2024-12-08T03:46:12,129 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=90, OpenFileDescriptor=439, MaxFileDescriptor=1048576, SystemLoadAverage=141, ProcessCount=11, AvailableMemoryMB=17505 2024-12-08T03:46:12,129 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.log.dir so I do NOT create it in target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/778c52cd-714d-d0cc-d695-8e602c72eba9/hadoop.tmp.dir so I do NOT create it in target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d, deleteOnExit=true 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/test.cache.data in system properties and HBase conf 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.tmp.dir in system properties and HBase conf 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir in system properties and HBase conf 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-08T03:46:12,130 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-08T03:46:12,130 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-08T03:46:12,131 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-08T03:46:12,131 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-08T03:46:12,131 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-08T03:46:12,131 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-08T03:46:12,131 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-08T03:46:12,131 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-08T03:46:12,131 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-08T03:46:12,132 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-08T03:46:12,132 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-08T03:46:12,132 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/nfs.dump.dir in system properties and HBase conf 2024-12-08T03:46:12,132 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/java.io.tmpdir in system properties and HBase conf 2024-12-08T03:46:12,132 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-08T03:46:12,132 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-08T03:46:12,132 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-08T03:46:12,396 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:12,401 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-08T03:46:12,402 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-08T03:46:12,402 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-08T03:46:12,402 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-08T03:46:12,403 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:12,403 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@377873f6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir/,AVAILABLE} 2024-12-08T03:46:12,404 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3950f25b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-08T03:46:12,492 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4f148449{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/java.io.tmpdir/jetty-localhost-45185-hadoop-hdfs-3_4_1-tests_jar-_-any-10832336199046519973/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-08T03:46:12,493 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4abfe850{HTTP/1.1, (http/1.1)}{localhost:45185} 2024-12-08T03:46:12,493 INFO [Time-limited test {}] server.Server(415): Started @10938ms 2024-12-08T03:46:12,704 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:12,707 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-08T03:46:12,708 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-08T03:46:12,708 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-08T03:46:12,708 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-08T03:46:12,708 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2aa8c76d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir/,AVAILABLE} 2024-12-08T03:46:12,709 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3ae83c3f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-08T03:46:12,798 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@12529ca{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/java.io.tmpdir/jetty-localhost-41901-hadoop-hdfs-3_4_1-tests_jar-_-any-12694539041532419699/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:12,799 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@847d000{HTTP/1.1, (http/1.1)}{localhost:41901} 2024-12-08T03:46:12,799 INFO [Time-limited test {}] server.Server(415): Started @11244ms 2024-12-08T03:46:12,800 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-08T03:46:12,826 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:12,829 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-08T03:46:12,831 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-08T03:46:12,831 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-08T03:46:12,831 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-08T03:46:12,832 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7ddf5a9c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir/,AVAILABLE} 2024-12-08T03:46:12,832 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@143655f2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-08T03:46:12,925 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1c5541de{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/java.io.tmpdir/jetty-localhost-34383-hadoop-hdfs-3_4_1-tests_jar-_-any-8105876643588737665/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:12,926 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@624e70a4{HTTP/1.1, (http/1.1)}{localhost:34383} 2024-12-08T03:46:12,926 INFO [Time-limited test {}] server.Server(415): Started @11371ms 2024-12-08T03:46:12,927 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-08T03:46:12,955 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-08T03:46:12,958 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-08T03:46:12,959 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-08T03:46:12,959 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-08T03:46:12,959 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-08T03:46:12,959 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@69168ef4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir/,AVAILABLE} 2024-12-08T03:46:12,959 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@659fda5a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-08T03:46:13,049 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@611d8bdf{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/java.io.tmpdir/jetty-localhost-37441-hadoop-hdfs-3_4_1-tests_jar-_-any-5883439607496675252/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:13,050 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@8e73203{HTTP/1.1, (http/1.1)}{localhost:37441} 2024-12-08T03:46:13,050 INFO [Time-limited test {}] server.Server(415): Started @11495ms 2024-12-08T03:46:13,052 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-08T03:46:13,626 WARN [Thread-558 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data1/current/BP-1091900809-172.17.0.2-1733629572156/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:13,626 WARN [Thread-559 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data2/current/BP-1091900809-172.17.0.2-1733629572156/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:13,639 WARN [Thread-500 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-08T03:46:13,642 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe29260f0bc112480 with lease ID 0xdb615108f4176161: Processing first storage report for DS-0b7cbd95-99d6-4027-bd2f-c1e332f72ca6 from datanode DatanodeRegistration(127.0.0.1:37631, datanodeUuid=322004e8-176c-4f0a-899b-eaab4d57e321, infoPort=34641, infoSecurePort=0, ipcPort=40003, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156) 2024-12-08T03:46:13,642 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe29260f0bc112480 with lease ID 0xdb615108f4176161: from storage DS-0b7cbd95-99d6-4027-bd2f-c1e332f72ca6 node DatanodeRegistration(127.0.0.1:37631, datanodeUuid=322004e8-176c-4f0a-899b-eaab4d57e321, infoPort=34641, infoSecurePort=0, ipcPort=40003, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:13,642 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe29260f0bc112480 with lease ID 0xdb615108f4176161: Processing first storage report for DS-786fd0c5-8326-4919-9b86-fa2cf6368786 from datanode DatanodeRegistration(127.0.0.1:37631, datanodeUuid=322004e8-176c-4f0a-899b-eaab4d57e321, infoPort=34641, infoSecurePort=0, ipcPort=40003, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156) 2024-12-08T03:46:13,642 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe29260f0bc112480 with lease ID 0xdb615108f4176161: from storage DS-786fd0c5-8326-4919-9b86-fa2cf6368786 node DatanodeRegistration(127.0.0.1:37631, datanodeUuid=322004e8-176c-4f0a-899b-eaab4d57e321, infoPort=34641, infoSecurePort=0, ipcPort=40003, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:13,993 WARN [Thread-571 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data3/current/BP-1091900809-172.17.0.2-1733629572156/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:13,993 WARN [Thread-572 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data4/current/BP-1091900809-172.17.0.2-1733629572156/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:14,012 WARN [Thread-523 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-08T03:46:14,017 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2aaf853a679d7807 with lease ID 0xdb615108f4176162: Processing first storage report for DS-4b033de5-a48f-4745-808d-1aa9c78b172f from datanode DatanodeRegistration(127.0.0.1:32999, datanodeUuid=298bfc20-1fd4-4c55-b99d-b422c5e9632f, infoPort=37917, infoSecurePort=0, ipcPort=39767, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156) 2024-12-08T03:46:14,017 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2aaf853a679d7807 with lease ID 0xdb615108f4176162: from storage DS-4b033de5-a48f-4745-808d-1aa9c78b172f node DatanodeRegistration(127.0.0.1:32999, datanodeUuid=298bfc20-1fd4-4c55-b99d-b422c5e9632f, infoPort=37917, infoSecurePort=0, ipcPort=39767, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:14,017 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2aaf853a679d7807 with lease ID 0xdb615108f4176162: Processing first storage report for DS-900ff7fe-63ad-40d9-81d8-5c6081776f4c from datanode DatanodeRegistration(127.0.0.1:32999, datanodeUuid=298bfc20-1fd4-4c55-b99d-b422c5e9632f, infoPort=37917, infoSecurePort=0, ipcPort=39767, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156) 2024-12-08T03:46:14,017 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2aaf853a679d7807 with lease ID 0xdb615108f4176162: from storage DS-900ff7fe-63ad-40d9-81d8-5c6081776f4c node DatanodeRegistration(127.0.0.1:32999, datanodeUuid=298bfc20-1fd4-4c55-b99d-b422c5e9632f, infoPort=37917, infoSecurePort=0, ipcPort=39767, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:14,037 WARN [Thread-582 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data5/current/BP-1091900809-172.17.0.2-1733629572156/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:14,037 WARN [Thread-583 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data6/current/BP-1091900809-172.17.0.2-1733629572156/current, will proceed with Du for space computation calculation, 2024-12-08T03:46:14,057 WARN [Thread-545 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-08T03:46:14,061 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xae26cea471eee2a with lease ID 0xdb615108f4176163: Processing first storage report for DS-8af72f4e-e8b7-42c1-ab7f-638394e21bd4 from datanode DatanodeRegistration(127.0.0.1:32941, datanodeUuid=a9b549fc-4f7d-4092-9b79-c1d7410b62b2, infoPort=45783, infoSecurePort=0, ipcPort=34625, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156) 2024-12-08T03:46:14,062 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xae26cea471eee2a with lease ID 0xdb615108f4176163: from storage DS-8af72f4e-e8b7-42c1-ab7f-638394e21bd4 node DatanodeRegistration(127.0.0.1:32941, datanodeUuid=a9b549fc-4f7d-4092-9b79-c1d7410b62b2, infoPort=45783, infoSecurePort=0, ipcPort=34625, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:14,062 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xae26cea471eee2a with lease ID 0xdb615108f4176163: Processing first storage report for DS-d1548002-140e-4605-a1c2-f0484f27def9 from datanode DatanodeRegistration(127.0.0.1:32941, datanodeUuid=a9b549fc-4f7d-4092-9b79-c1d7410b62b2, infoPort=45783, infoSecurePort=0, ipcPort=34625, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156) 2024-12-08T03:46:14,062 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xae26cea471eee2a with lease ID 0xdb615108f4176163: from storage DS-d1548002-140e-4605-a1c2-f0484f27def9 node DatanodeRegistration(127.0.0.1:32941, datanodeUuid=a9b549fc-4f7d-4092-9b79-c1d7410b62b2, infoPort=45783, infoSecurePort=0, ipcPort=34625, storageInfo=lv=-57;cid=testClusterID;nsid=808664083;c=1733629572156), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-08T03:46:14,096 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a 2024-12-08T03:46:14,099 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/zookeeper_0, clientPort=53394, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-08T03:46:14,100 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=53394 2024-12-08T03:46:14,100 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,101 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,114 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741825_1001 (size=7) 2024-12-08T03:46:14,115 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741825_1001 (size=7) 2024-12-08T03:46:14,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741825_1001 (size=7) 2024-12-08T03:46:14,117 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe with version=8 2024-12-08T03:46:14,117 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:42793/user/jenkins/test-data/733738b7-1f32-8392-cf5c-0c611b0b6791/hbase-staging 2024-12-08T03:46:14,120 INFO [Time-limited test {}] client.ConnectionUtils(128): master/32df4b4bb439:0 server-side Connection retries=45 2024-12-08T03:46:14,120 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,120 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,120 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-08T03:46:14,120 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,120 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-08T03:46:14,120 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-08T03:46:14,120 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-08T03:46:14,121 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:37713 2024-12-08T03:46:14,122 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:37713 connecting to ZooKeeper ensemble=127.0.0.1:53394 2024-12-08T03:46:14,169 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:377130x0, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-08T03:46:14,170 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:37713-0x1000395ba6f0000 connected 2024-12-08T03:46:14,249 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,252 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,255 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:14,255 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe, hbase.cluster.distributed=false 2024-12-08T03:46:14,259 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-08T03:46:14,259 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37713 2024-12-08T03:46:14,260 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37713 2024-12-08T03:46:14,260 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37713 2024-12-08T03:46:14,261 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37713 2024-12-08T03:46:14,262 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37713 2024-12-08T03:46:14,279 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32df4b4bb439:0 server-side Connection retries=45 2024-12-08T03:46:14,279 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,279 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,279 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-08T03:46:14,279 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,279 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-08T03:46:14,279 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-08T03:46:14,280 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-08T03:46:14,280 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:43979 2024-12-08T03:46:14,281 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:43979 connecting to ZooKeeper ensemble=127.0.0.1:53394 2024-12-08T03:46:14,282 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,284 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,297 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:439790x0, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-08T03:46:14,297 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:43979-0x1000395ba6f0001 connected 2024-12-08T03:46:14,297 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:14,298 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-08T03:46:14,298 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-08T03:46:14,299 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-08T03:46:14,300 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-08T03:46:14,301 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43979 2024-12-08T03:46:14,301 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43979 2024-12-08T03:46:14,302 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43979 2024-12-08T03:46:14,302 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43979 2024-12-08T03:46:14,302 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43979 2024-12-08T03:46:14,319 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32df4b4bb439:0 server-side Connection retries=45 2024-12-08T03:46:14,319 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,319 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,319 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-08T03:46:14,319 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,319 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-08T03:46:14,319 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-08T03:46:14,319 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-08T03:46:14,320 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:34905 2024-12-08T03:46:14,321 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:34905 connecting to ZooKeeper ensemble=127.0.0.1:53394 2024-12-08T03:46:14,322 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,323 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,332 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:349050x0, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-08T03:46:14,333 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:34905-0x1000395ba6f0002 connected 2024-12-08T03:46:14,333 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:14,333 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-08T03:46:14,333 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-08T03:46:14,334 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-08T03:46:14,335 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-08T03:46:14,336 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34905 2024-12-08T03:46:14,336 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34905 2024-12-08T03:46:14,336 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34905 2024-12-08T03:46:14,337 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34905 2024-12-08T03:46:14,337 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34905 2024-12-08T03:46:14,353 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32df4b4bb439:0 server-side Connection retries=45 2024-12-08T03:46:14,353 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,353 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,353 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-08T03:46:14,353 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-08T03:46:14,353 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-08T03:46:14,353 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-08T03:46:14,353 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-08T03:46:14,354 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:44085 2024-12-08T03:46:14,355 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:44085 connecting to ZooKeeper ensemble=127.0.0.1:53394 2024-12-08T03:46:14,356 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,357 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:440850x0, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-08T03:46:14,366 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:44085-0x1000395ba6f0003 connected 2024-12-08T03:46:14,366 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:14,366 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-08T03:46:14,367 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-08T03:46:14,368 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-08T03:46:14,369 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-08T03:46:14,369 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=44085 2024-12-08T03:46:14,369 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=44085 2024-12-08T03:46:14,370 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=44085 2024-12-08T03:46:14,370 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=44085 2024-12-08T03:46:14,370 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=44085 2024-12-08T03:46:14,383 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;32df4b4bb439:37713 2024-12-08T03:46:14,383 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/32df4b4bb439,37713,1733629574119 2024-12-08T03:46:14,390 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:14,390 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:14,390 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:14,390 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:14,391 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/32df4b4bb439,37713,1733629574119 2024-12-08T03:46:14,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-08T03:46:14,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-08T03:46:14,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-08T03:46:14,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,400 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-08T03:46:14,400 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/32df4b4bb439,37713,1733629574119 from backup master directory 2024-12-08T03:46:14,407 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:14,407 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:14,407 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:14,407 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/32df4b4bb439,37713,1733629574119 2024-12-08T03:46:14,407 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-08T03:46:14,407 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-08T03:46:14,407 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=32df4b4bb439,37713,1733629574119 2024-12-08T03:46:14,413 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/hbase.id] with ID: 0a486559-1124-430f-8e60-95066459bb19 2024-12-08T03:46:14,413 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/.tmp/hbase.id 2024-12-08T03:46:14,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741826_1002 (size=42) 2024-12-08T03:46:14,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741826_1002 (size=42) 2024-12-08T03:46:14,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741826_1002 (size=42) 2024-12-08T03:46:14,423 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/.tmp/hbase.id]:[hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/hbase.id] 2024-12-08T03:46:14,439 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-08T03:46:14,439 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-08T03:46:14,440 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 1ms. 2024-12-08T03:46:14,449 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,449 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,449 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,449 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,457 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741827_1003 (size=196) 2024-12-08T03:46:14,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741827_1003 (size=196) 2024-12-08T03:46:14,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741827_1003 (size=196) 2024-12-08T03:46:14,459 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-08T03:46:14,459 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-08T03:46:14,460 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-08T03:46:14,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741828_1004 (size=1189) 2024-12-08T03:46:14,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741828_1004 (size=1189) 2024-12-08T03:46:14,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741828_1004 (size=1189) 2024-12-08T03:46:14,472 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store 2024-12-08T03:46:14,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741829_1005 (size=34) 2024-12-08T03:46:14,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741829_1005 (size=34) 2024-12-08T03:46:14,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741829_1005 (size=34) 2024-12-08T03:46:14,484 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:14,484 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-08T03:46:14,484 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:14,484 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:14,484 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-08T03:46:14,484 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:14,484 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:14,484 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733629574484Disabling compacts and flushes for region at 1733629574484Disabling writes for close at 1733629574484Writing region close event to WAL at 1733629574484Closed at 1733629574484 2024-12-08T03:46:14,486 WARN [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/.initializing 2024-12-08T03:46:14,486 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/WALs/32df4b4bb439,37713,1733629574119 2024-12-08T03:46:14,490 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C37713%2C1733629574119, suffix=, logDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/WALs/32df4b4bb439,37713,1733629574119, archiveDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/oldWALs, maxLogs=10 2024-12-08T03:46:14,491 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 32df4b4bb439%2C37713%2C1733629574119.1733629574490 2024-12-08T03:46:14,501 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/WALs/32df4b4bb439,37713,1733629574119/32df4b4bb439%2C37713%2C1733629574119.1733629574490 2024-12-08T03:46:14,507 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45783:45783),(127.0.0.1/127.0.0.1:34641:34641),(127.0.0.1/127.0.0.1:37917:37917)] 2024-12-08T03:46:14,507 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-08T03:46:14,508 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:14,508 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,508 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,510 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,512 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-08T03:46:14,512 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,513 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:14,513 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,515 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-08T03:46:14,515 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,515 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-08T03:46:14,516 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,518 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-08T03:46:14,518 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,519 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-08T03:46:14,519 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,521 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-08T03:46:14,521 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,522 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-08T03:46:14,522 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,523 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,523 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,525 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,525 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,525 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-08T03:46:14,527 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-08T03:46:14,529 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-08T03:46:14,530 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69276655, jitterRate=0.032302603125572205}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-08T03:46:14,530 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733629574508Initializing all the Stores at 1733629574509 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629574509Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629574510 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629574510Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629574510Cleaning up temporary data from old regions at 1733629574525 (+15 ms)Region opened successfully at 1733629574530 (+5 ms) 2024-12-08T03:46:14,531 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-08T03:46:14,535 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7d21089e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32df4b4bb439/172.17.0.2:0 2024-12-08T03:46:14,536 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-08T03:46:14,536 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-08T03:46:14,536 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-08T03:46:14,536 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-08T03:46:14,537 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-08T03:46:14,537 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-08T03:46:14,538 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-08T03:46:14,540 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-08T03:46:14,541 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-08T03:46:14,557 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-08T03:46:14,558 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-08T03:46:14,558 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-08T03:46:14,565 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-08T03:46:14,566 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-08T03:46:14,567 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-08T03:46:14,574 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-08T03:46:14,575 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-08T03:46:14,582 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-08T03:46:14,584 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-08T03:46:14,590 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-08T03:46:14,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:14,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:14,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:14,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:14,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,599 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,599 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=32df4b4bb439,37713,1733629574119, sessionid=0x1000395ba6f0000, setting cluster-up flag (Was=false) 2024-12-08T03:46:14,615 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,615 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,615 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,615 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,640 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-08T03:46:14,642 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=32df4b4bb439,37713,1733629574119 2024-12-08T03:46:14,666 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,666 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,666 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,666 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:14,690 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-08T03:46:14,692 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=32df4b4bb439,37713,1733629574119 2024-12-08T03:46:14,693 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-08T03:46:14,695 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-08T03:46:14,696 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-08T03:46:14,696 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-08T03:46:14,696 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 32df4b4bb439,37713,1733629574119 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-08T03:46:14,698 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/32df4b4bb439:0, corePoolSize=5, maxPoolSize=5 2024-12-08T03:46:14,698 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/32df4b4bb439:0, corePoolSize=5, maxPoolSize=5 2024-12-08T03:46:14,698 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/32df4b4bb439:0, corePoolSize=5, maxPoolSize=5 2024-12-08T03:46:14,698 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/32df4b4bb439:0, corePoolSize=5, maxPoolSize=5 2024-12-08T03:46:14,698 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/32df4b4bb439:0, corePoolSize=10, maxPoolSize=10 2024-12-08T03:46:14,698 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,698 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/32df4b4bb439:0, corePoolSize=2, maxPoolSize=2 2024-12-08T03:46:14,698 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,700 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733629604700 2024-12-08T03:46:14,700 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-08T03:46:14,700 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-08T03:46:14,700 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-08T03:46:14,700 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-08T03:46:14,700 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-08T03:46:14,700 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-08T03:46:14,700 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,701 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-08T03:46:14,701 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-08T03:46:14,701 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-08T03:46:14,701 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-08T03:46:14,701 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-08T03:46:14,701 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-08T03:46:14,701 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-08T03:46:14,701 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.large.0-1733629574701,5,FailOnTimeoutGroup] 2024-12-08T03:46:14,702 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,702 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-08T03:46:14,704 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.small.0-1733629574702,5,FailOnTimeoutGroup] 2024-12-08T03:46:14,704 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,704 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-08T03:46:14,704 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,704 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741831_1007 (size=1321) 2024-12-08T03:46:14,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741831_1007 (size=1321) 2024-12-08T03:46:14,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741831_1007 (size=1321) 2024-12-08T03:46:14,713 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-08T03:46:14,713 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe 2024-12-08T03:46:14,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741832_1008 (size=32) 2024-12-08T03:46:14,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741832_1008 (size=32) 2024-12-08T03:46:14,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741832_1008 (size=32) 2024-12-08T03:46:14,723 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:14,725 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-08T03:46:14,726 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-08T03:46:14,726 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,727 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:14,727 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-08T03:46:14,728 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-08T03:46:14,728 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,728 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:14,728 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-08T03:46:14,730 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-08T03:46:14,730 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,730 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:14,730 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-08T03:46:14,731 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-08T03:46:14,732 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:14,732 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:14,732 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-08T03:46:14,733 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740 2024-12-08T03:46:14,733 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740 2024-12-08T03:46:14,735 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-08T03:46:14,735 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-08T03:46:14,735 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-08T03:46:14,737 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-08T03:46:14,739 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-08T03:46:14,740 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71215567, jitterRate=0.06119464337825775}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-08T03:46:14,740 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733629574723Initializing all the Stores at 1733629574724 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629574724Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629574724Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629574724Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629574724Cleaning up temporary data from old regions at 1733629574735 (+11 ms)Region opened successfully at 1733629574740 (+5 ms) 2024-12-08T03:46:14,741 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-08T03:46:14,741 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-08T03:46:14,741 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-08T03:46:14,741 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-08T03:46:14,741 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-08T03:46:14,741 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-08T03:46:14,741 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733629574741Disabling compacts and flushes for region at 1733629574741Disabling writes for close at 1733629574741Writing region close event to WAL at 1733629574741Closed at 1733629574741 2024-12-08T03:46:14,743 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-08T03:46:14,743 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-08T03:46:14,743 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-08T03:46:14,745 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-08T03:46:14,746 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-08T03:46:14,772 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(746): ClusterId : 0a486559-1124-430f-8e60-95066459bb19 2024-12-08T03:46:14,772 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(746): ClusterId : 0a486559-1124-430f-8e60-95066459bb19 2024-12-08T03:46:14,772 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(746): ClusterId : 0a486559-1124-430f-8e60-95066459bb19 2024-12-08T03:46:14,772 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-08T03:46:14,772 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-08T03:46:14,772 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-08T03:46:14,800 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-08T03:46:14,800 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-08T03:46:14,800 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-08T03:46:14,800 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-08T03:46:14,800 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-08T03:46:14,800 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-08T03:46:14,808 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-08T03:46:14,808 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-08T03:46:14,808 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-08T03:46:14,809 DEBUG [RS:1;32df4b4bb439:34905 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@64943e50, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32df4b4bb439/172.17.0.2:0 2024-12-08T03:46:14,809 DEBUG [RS:2;32df4b4bb439:44085 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@58b305fc, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32df4b4bb439/172.17.0.2:0 2024-12-08T03:46:14,809 DEBUG [RS:0;32df4b4bb439:43979 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@551384cd, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32df4b4bb439/172.17.0.2:0 2024-12-08T03:46:14,822 DEBUG [RS:2;32df4b4bb439:44085 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;32df4b4bb439:44085 2024-12-08T03:46:14,822 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-08T03:46:14,822 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-08T03:46:14,822 DEBUG [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-08T03:46:14,824 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(2659): reportForDuty to master=32df4b4bb439,37713,1733629574119 with port=44085, startcode=1733629574352 2024-12-08T03:46:14,824 DEBUG [RS:2;32df4b4bb439:44085 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-08T03:46:14,825 DEBUG [RS:1;32df4b4bb439:34905 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;32df4b4bb439:34905 2024-12-08T03:46:14,825 DEBUG [RS:0;32df4b4bb439:43979 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;32df4b4bb439:43979 2024-12-08T03:46:14,825 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-08T03:46:14,825 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-08T03:46:14,825 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-08T03:46:14,825 DEBUG [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-08T03:46:14,825 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-08T03:46:14,825 DEBUG [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-08T03:46:14,826 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52879, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-08T03:46:14,826 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(2659): reportForDuty to master=32df4b4bb439,37713,1733629574119 with port=34905, startcode=1733629574318 2024-12-08T03:46:14,826 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(2659): reportForDuty to master=32df4b4bb439,37713,1733629574119 with port=43979, startcode=1733629574279 2024-12-08T03:46:14,826 DEBUG [RS:1;32df4b4bb439:34905 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-08T03:46:14,826 DEBUG [RS:0;32df4b4bb439:43979 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-08T03:46:14,827 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37713 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32df4b4bb439,44085,1733629574352 2024-12-08T03:46:14,827 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37713 {}] master.ServerManager(517): Registering regionserver=32df4b4bb439,44085,1733629574352 2024-12-08T03:46:14,829 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:38051, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-08T03:46:14,829 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41361, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-08T03:46:14,830 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37713 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32df4b4bb439,34905,1733629574318 2024-12-08T03:46:14,830 DEBUG [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe 2024-12-08T03:46:14,830 DEBUG [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41623 2024-12-08T03:46:14,830 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37713 {}] master.ServerManager(517): Registering regionserver=32df4b4bb439,34905,1733629574318 2024-12-08T03:46:14,830 DEBUG [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-08T03:46:14,832 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37713 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32df4b4bb439,43979,1733629574279 2024-12-08T03:46:14,832 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37713 {}] master.ServerManager(517): Registering regionserver=32df4b4bb439,43979,1733629574279 2024-12-08T03:46:14,832 DEBUG [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe 2024-12-08T03:46:14,832 DEBUG [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41623 2024-12-08T03:46:14,832 DEBUG [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-08T03:46:14,834 DEBUG [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe 2024-12-08T03:46:14,834 DEBUG [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41623 2024-12-08T03:46:14,834 DEBUG [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-08T03:46:14,840 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-08T03:46:14,869 DEBUG [RS:2;32df4b4bb439:44085 {}] zookeeper.ZKUtil(111): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32df4b4bb439,44085,1733629574352 2024-12-08T03:46:14,869 WARN [RS:2;32df4b4bb439:44085 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-08T03:46:14,869 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32df4b4bb439,43979,1733629574279] 2024-12-08T03:46:14,869 INFO [RS:2;32df4b4bb439:44085 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-08T03:46:14,869 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32df4b4bb439,44085,1733629574352] 2024-12-08T03:46:14,869 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32df4b4bb439,34905,1733629574318] 2024-12-08T03:46:14,870 DEBUG [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,44085,1733629574352 2024-12-08T03:46:14,870 DEBUG [RS:1;32df4b4bb439:34905 {}] zookeeper.ZKUtil(111): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32df4b4bb439,34905,1733629574318 2024-12-08T03:46:14,870 DEBUG [RS:0;32df4b4bb439:43979 {}] zookeeper.ZKUtil(111): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32df4b4bb439,43979,1733629574279 2024-12-08T03:46:14,870 WARN [RS:1;32df4b4bb439:34905 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-08T03:46:14,870 WARN [RS:0;32df4b4bb439:43979 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-08T03:46:14,870 INFO [RS:1;32df4b4bb439:34905 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-08T03:46:14,870 INFO [RS:0;32df4b4bb439:43979 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-08T03:46:14,870 DEBUG [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,43979,1733629574279 2024-12-08T03:46:14,870 DEBUG [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,34905,1733629574318 2024-12-08T03:46:14,876 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-08T03:46:14,876 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-08T03:46:14,876 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-08T03:46:14,879 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-08T03:46:14,880 INFO [RS:2;32df4b4bb439:44085 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-08T03:46:14,880 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,880 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-08T03:46:14,881 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-08T03:46:14,881 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,881 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32df4b4bb439:0, corePoolSize=2, maxPoolSize=2 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,882 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,883 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:14,883 DEBUG [RS:2;32df4b4bb439:44085 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:14,883 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-08T03:46:14,883 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-08T03:46:14,883 INFO [RS:1;32df4b4bb439:34905 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-08T03:46:14,883 INFO [RS:0;32df4b4bb439:43979 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-08T03:46:14,883 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,883 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,884 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,884 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,884 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-08T03:46:14,884 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,884 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,884 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,884 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,44085,1733629574352-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-08T03:46:14,885 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-08T03:46:14,885 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32df4b4bb439:0, corePoolSize=2, maxPoolSize=2 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,885 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,886 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:14,886 DEBUG [RS:0;32df4b4bb439:43979 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:14,887 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-08T03:46:14,888 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-08T03:46:14,888 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,888 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,888 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,888 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,888 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,888 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,888 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,888 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,888 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,43979,1733629574279-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-08T03:46:14,888 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,888 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,888 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,888 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32df4b4bb439:0, corePoolSize=2, maxPoolSize=2 2024-12-08T03:46:14,888 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,889 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,889 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,889 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,889 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,889 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32df4b4bb439:0, corePoolSize=1, maxPoolSize=1 2024-12-08T03:46:14,889 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:14,889 DEBUG [RS:1;32df4b4bb439:34905 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0, corePoolSize=3, maxPoolSize=3 2024-12-08T03:46:14,892 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,892 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,893 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,893 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,893 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,893 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,34905,1733629574318-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-08T03:46:14,897 WARN [32df4b4bb439:37713 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-08T03:46:14,901 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-08T03:46:14,901 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,44085,1733629574352-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,901 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,901 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.Replication(171): 32df4b4bb439,44085,1733629574352 started 2024-12-08T03:46:14,904 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-08T03:46:14,904 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-08T03:46:14,905 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,34905,1733629574318-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,905 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,43979,1733629574279-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,905 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,905 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,905 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.Replication(171): 32df4b4bb439,43979,1733629574279 started 2024-12-08T03:46:14,905 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.Replication(171): 32df4b4bb439,34905,1733629574318 started 2024-12-08T03:46:14,913 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,913 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1482): Serving as 32df4b4bb439,44085,1733629574352, RpcServer on 32df4b4bb439/172.17.0.2:44085, sessionid=0x1000395ba6f0003 2024-12-08T03:46:14,913 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-08T03:46:14,913 DEBUG [RS:2;32df4b4bb439:44085 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32df4b4bb439,44085,1733629574352 2024-12-08T03:46:14,914 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,44085,1733629574352' 2024-12-08T03:46:14,914 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-08T03:46:14,914 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-08T03:46:14,915 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-08T03:46:14,915 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-08T03:46:14,915 DEBUG [RS:2;32df4b4bb439:44085 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32df4b4bb439,44085,1733629574352 2024-12-08T03:46:14,915 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,44085,1733629574352' 2024-12-08T03:46:14,915 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-08T03:46:14,915 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-08T03:46:14,916 DEBUG [RS:2;32df4b4bb439:44085 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-08T03:46:14,916 INFO [RS:2;32df4b4bb439:44085 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-08T03:46:14,916 INFO [RS:2;32df4b4bb439:44085 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-08T03:46:14,916 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,916 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(1482): Serving as 32df4b4bb439,43979,1733629574279, RpcServer on 32df4b4bb439/172.17.0.2:43979, sessionid=0x1000395ba6f0001 2024-12-08T03:46:14,916 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:14,917 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(1482): Serving as 32df4b4bb439,34905,1733629574318, RpcServer on 32df4b4bb439/172.17.0.2:34905, sessionid=0x1000395ba6f0002 2024-12-08T03:46:14,917 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-08T03:46:14,917 DEBUG [RS:0;32df4b4bb439:43979 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32df4b4bb439,43979,1733629574279 2024-12-08T03:46:14,917 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,43979,1733629574279' 2024-12-08T03:46:14,917 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-08T03:46:14,917 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-08T03:46:14,917 DEBUG [RS:1;32df4b4bb439:34905 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32df4b4bb439,34905,1733629574318 2024-12-08T03:46:14,917 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,34905,1733629574318' 2024-12-08T03:46:14,917 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-08T03:46:14,917 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-08T03:46:14,917 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-08T03:46:14,918 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-08T03:46:14,918 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-08T03:46:14,918 DEBUG [RS:0;32df4b4bb439:43979 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32df4b4bb439,43979,1733629574279 2024-12-08T03:46:14,918 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-08T03:46:14,918 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,43979,1733629574279' 2024-12-08T03:46:14,918 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-08T03:46:14,918 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-08T03:46:14,918 DEBUG [RS:1;32df4b4bb439:34905 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32df4b4bb439,34905,1733629574318 2024-12-08T03:46:14,918 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32df4b4bb439,34905,1733629574318' 2024-12-08T03:46:14,918 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-08T03:46:14,918 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-08T03:46:14,918 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-08T03:46:14,919 DEBUG [RS:0;32df4b4bb439:43979 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-08T03:46:14,919 INFO [RS:0;32df4b4bb439:43979 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-08T03:46:14,919 INFO [RS:0;32df4b4bb439:43979 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-08T03:46:14,919 DEBUG [RS:1;32df4b4bb439:34905 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-08T03:46:14,919 INFO [RS:1;32df4b4bb439:34905 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-08T03:46:14,919 INFO [RS:1;32df4b4bb439:34905 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-08T03:46:15,020 INFO [RS:2;32df4b4bb439:44085 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C44085%2C1733629574352, suffix=, logDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,44085,1733629574352, archiveDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/oldWALs, maxLogs=32 2024-12-08T03:46:15,023 INFO [RS:0;32df4b4bb439:43979 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C43979%2C1733629574279, suffix=, logDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,43979,1733629574279, archiveDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/oldWALs, maxLogs=32 2024-12-08T03:46:15,023 INFO [RS:1;32df4b4bb439:34905 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C34905%2C1733629574318, suffix=, logDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,34905,1733629574318, archiveDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/oldWALs, maxLogs=32 2024-12-08T03:46:15,026 INFO [RS:2;32df4b4bb439:44085 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 32df4b4bb439%2C44085%2C1733629574352.1733629575025 2024-12-08T03:46:15,027 INFO [RS:1;32df4b4bb439:34905 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 32df4b4bb439%2C34905%2C1733629574318.1733629575027 2024-12-08T03:46:15,027 INFO [RS:0;32df4b4bb439:43979 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 32df4b4bb439%2C43979%2C1733629574279.1733629575027 2024-12-08T03:46:15,039 INFO [RS:2;32df4b4bb439:44085 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,44085,1733629574352/32df4b4bb439%2C44085%2C1733629574352.1733629575025 2024-12-08T03:46:15,043 INFO [RS:0;32df4b4bb439:43979 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,43979,1733629574279/32df4b4bb439%2C43979%2C1733629574279.1733629575027 2024-12-08T03:46:15,043 INFO [RS:1;32df4b4bb439:34905 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,34905,1733629574318/32df4b4bb439%2C34905%2C1733629574318.1733629575027 2024-12-08T03:46:15,045 DEBUG [RS:2;32df4b4bb439:44085 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45783:45783),(127.0.0.1/127.0.0.1:37917:37917),(127.0.0.1/127.0.0.1:34641:34641)] 2024-12-08T03:46:15,046 DEBUG [RS:0;32df4b4bb439:43979 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45783:45783),(127.0.0.1/127.0.0.1:37917:37917),(127.0.0.1/127.0.0.1:34641:34641)] 2024-12-08T03:46:15,046 DEBUG [RS:1;32df4b4bb439:34905 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34641:34641),(127.0.0.1/127.0.0.1:37917:37917),(127.0.0.1/127.0.0.1:45783:45783)] 2024-12-08T03:46:15,147 DEBUG [32df4b4bb439:37713 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-08T03:46:15,148 DEBUG [32df4b4bb439:37713 {}] balancer.BalancerClusterState(204): Hosts are {32df4b4bb439=0} racks are {/default-rack=0} 2024-12-08T03:46:15,155 DEBUG [32df4b4bb439:37713 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-08T03:46:15,155 DEBUG [32df4b4bb439:37713 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-08T03:46:15,155 DEBUG [32df4b4bb439:37713 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-08T03:46:15,155 DEBUG [32df4b4bb439:37713 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-08T03:46:15,155 DEBUG [32df4b4bb439:37713 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-08T03:46:15,155 DEBUG [32df4b4bb439:37713 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-08T03:46:15,155 INFO [32df4b4bb439:37713 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-08T03:46:15,155 INFO [32df4b4bb439:37713 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-08T03:46:15,155 INFO [32df4b4bb439:37713 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-08T03:46:15,156 DEBUG [32df4b4bb439:37713 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-08T03:46:15,156 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=32df4b4bb439,44085,1733629574352 2024-12-08T03:46:15,158 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 32df4b4bb439,44085,1733629574352, state=OPENING 2024-12-08T03:46:15,190 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-08T03:46:15,199 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:15,199 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:15,199 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:15,199 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:15,200 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-08T03:46:15,200 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:15,200 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:15,200 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:15,200 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=32df4b4bb439,44085,1733629574352}] 2024-12-08T03:46:15,200 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:15,357 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-08T03:46:15,361 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45597, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-08T03:46:15,368 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-08T03:46:15,369 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-08T03:46:15,371 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32df4b4bb439%2C44085%2C1733629574352.meta, suffix=.meta, logDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,44085,1733629574352, archiveDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/oldWALs, maxLogs=32 2024-12-08T03:46:15,372 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 32df4b4bb439%2C44085%2C1733629574352.meta.1733629575372.meta 2024-12-08T03:46:15,379 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/WALs/32df4b4bb439,44085,1733629574352/32df4b4bb439%2C44085%2C1733629574352.meta.1733629575372.meta 2024-12-08T03:46:15,381 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37917:37917),(127.0.0.1/127.0.0.1:34641:34641),(127.0.0.1/127.0.0.1:45783:45783)] 2024-12-08T03:46:15,383 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-08T03:46:15,384 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-08T03:46:15,384 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-08T03:46:15,384 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-08T03:46:15,384 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-08T03:46:15,384 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:15,384 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-08T03:46:15,384 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-08T03:46:15,386 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-08T03:46:15,387 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-08T03:46:15,387 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:15,388 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:15,388 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-08T03:46:15,389 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-08T03:46:15,389 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:15,389 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:15,389 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-08T03:46:15,390 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-08T03:46:15,390 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:15,391 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:15,391 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-08T03:46:15,391 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-08T03:46:15,392 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:15,392 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-08T03:46:15,392 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-08T03:46:15,393 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740 2024-12-08T03:46:15,395 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740 2024-12-08T03:46:15,396 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-08T03:46:15,396 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-08T03:46:15,397 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-08T03:46:15,398 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-08T03:46:15,398 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-08T03:46:15,399 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73650830, jitterRate=0.09748288989067078}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-08T03:46:15,399 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-08T03:46:15,401 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733629575385Writing region info on filesystem at 1733629575385Initializing all the Stores at 1733629575386 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629575386Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629575386Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629575386Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733629575386Cleaning up temporary data from old regions at 1733629575396 (+10 ms)Running coprocessor post-open hooks at 1733629575399 (+3 ms)Region opened successfully at 1733629575401 (+2 ms) 2024-12-08T03:46:15,403 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733629575357 2024-12-08T03:46:15,404 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-08T03:46:15,407 DEBUG [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-08T03:46:15,407 INFO [RS_OPEN_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-08T03:46:15,408 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=32df4b4bb439,44085,1733629574352 2024-12-08T03:46:15,410 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 32df4b4bb439,44085,1733629574352, state=OPEN 2024-12-08T03:46:15,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-08T03:46:15,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-08T03:46:15,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-08T03:46:15,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-08T03:46:15,432 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=32df4b4bb439,44085,1733629574352 2024-12-08T03:46:15,432 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:15,432 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:15,432 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:15,432 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-08T03:46:15,436 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-08T03:46:15,436 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=32df4b4bb439,44085,1733629574352 in 232 msec 2024-12-08T03:46:15,440 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-08T03:46:15,440 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 693 msec 2024-12-08T03:46:15,441 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-08T03:46:15,441 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-08T03:46:15,443 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-08T03:46:15,443 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=32df4b4bb439,44085,1733629574352, seqNum=-1] 2024-12-08T03:46:15,443 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-08T03:46:15,445 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41125, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-08T03:46:15,452 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 756 msec 2024-12-08T03:46:15,453 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733629575452, completionTime=-1 2024-12-08T03:46:15,453 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-08T03:46:15,453 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-08T03:46:15,455 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-08T03:46:15,455 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733629635455 2024-12-08T03:46:15,455 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733629695455 2024-12-08T03:46:15,455 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-12-08T03:46:15,455 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37713,1733629574119-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:15,455 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37713,1733629574119-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:15,455 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37713,1733629574119-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:15,456 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-32df4b4bb439:37713, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:15,456 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:15,456 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:15,458 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-08T03:46:15,458 DEBUG [master/32df4b4bb439:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-08T03:46:15,459 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-08T03:46:15,461 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.054sec 2024-12-08T03:46:15,461 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-08T03:46:15,461 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-08T03:46:15,461 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-08T03:46:15,461 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-08T03:46:15,462 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-08T03:46:15,462 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37713,1733629574119-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-08T03:46:15,462 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37713,1733629574119-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-08T03:46:15,464 DEBUG [master/32df4b4bb439:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-08T03:46:15,465 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-08T03:46:15,465 INFO [master/32df4b4bb439:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32df4b4bb439,37713,1733629574119-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-08T03:46:15,473 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@18223ed7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-08T03:46:15,473 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 32df4b4bb439,37713,-1 for getting cluster id 2024-12-08T03:46:15,473 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-08T03:46:15,474 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '0a486559-1124-430f-8e60-95066459bb19' 2024-12-08T03:46:15,474 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-08T03:46:15,474 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "0a486559-1124-430f-8e60-95066459bb19" 2024-12-08T03:46:15,475 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@25868db1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-08T03:46:15,475 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [32df4b4bb439,37713,-1] 2024-12-08T03:46:15,475 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-08T03:46:15,475 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:15,477 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42560, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-08T03:46:15,478 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@20a75a5c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-08T03:46:15,478 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-08T03:46:15,480 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=32df4b4bb439,44085,1733629574352, seqNum=-1] 2024-12-08T03:46:15,480 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-08T03:46:15,482 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60368, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-08T03:46:15,484 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=32df4b4bb439,37713,1733629574119 2024-12-08T03:46:15,485 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-08T03:46:15,486 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.AsyncConnectionImpl(321): The fetched master address is 32df4b4bb439,37713,1733629574119 2024-12-08T03:46:15,486 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@14a47ed0 2024-12-08T03:46:15,486 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-08T03:46:15,488 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42572, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-08T03:46:15,489 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-08T03:46:15,491 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-08T03:46:15,493 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-08T03:46:15,494 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:15,494 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-08T03:46:15,495 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-08T03:46:15,496 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-08T03:46:15,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741837_1013 (size=392) 2024-12-08T03:46:15,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741837_1013 (size=392) 2024-12-08T03:46:15,507 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741837_1013 (size=392) 2024-12-08T03:46:15,509 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => f9a74e4ccdae036154df7dfc04494f2b, NAME => 'TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe 2024-12-08T03:46:15,521 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741838_1014 (size=51) 2024-12-08T03:46:15,521 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741838_1014 (size=51) 2024-12-08T03:46:15,521 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741838_1014 (size=51) 2024-12-08T03:46:15,522 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:15,522 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing f9a74e4ccdae036154df7dfc04494f2b, disabling compactions & flushes 2024-12-08T03:46:15,522 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:15,522 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:15,522 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. after waiting 0 ms 2024-12-08T03:46:15,522 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:15,522 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:15,522 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for f9a74e4ccdae036154df7dfc04494f2b: Waiting for close lock at 1733629575522Disabling compacts and flushes for region at 1733629575522Disabling writes for close at 1733629575522Writing region close event to WAL at 1733629575522Closed at 1733629575522 2024-12-08T03:46:15,524 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-08T03:46:15,525 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733629575525"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733629575525"}]},"ts":"1733629575525"} 2024-12-08T03:46:15,528 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-08T03:46:15,530 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-08T03:46:15,530 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733629575530"}]},"ts":"1733629575530"} 2024-12-08T03:46:15,533 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-08T03:46:15,534 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {32df4b4bb439=0} racks are {/default-rack=0} 2024-12-08T03:46:15,535 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-08T03:46:15,535 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-08T03:46:15,535 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-08T03:46:15,535 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-08T03:46:15,535 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-08T03:46:15,535 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-08T03:46:15,535 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-08T03:46:15,535 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-08T03:46:15,535 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-08T03:46:15,535 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-08T03:46:15,535 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=f9a74e4ccdae036154df7dfc04494f2b, ASSIGN}] 2024-12-08T03:46:15,537 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=f9a74e4ccdae036154df7dfc04494f2b, ASSIGN 2024-12-08T03:46:15,539 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=f9a74e4ccdae036154df7dfc04494f2b, ASSIGN; state=OFFLINE, location=32df4b4bb439,44085,1733629574352; forceNewPlan=false, retain=false 2024-12-08T03:46:15,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-08T03:46:15,690 INFO [32df4b4bb439:37713 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-08T03:46:15,690 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=f9a74e4ccdae036154df7dfc04494f2b, regionState=OPENING, regionLocation=32df4b4bb439,44085,1733629574352 2024-12-08T03:46:15,694 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=f9a74e4ccdae036154df7dfc04494f2b, ASSIGN because future has completed 2024-12-08T03:46:15,694 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure f9a74e4ccdae036154df7dfc04494f2b, server=32df4b4bb439,44085,1733629574352}] 2024-12-08T03:46:15,812 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-08T03:46:15,857 INFO [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:15,858 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => f9a74e4ccdae036154df7dfc04494f2b, NAME => 'TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b.', STARTKEY => '', ENDKEY => ''} 2024-12-08T03:46:15,858 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,858 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-08T03:46:15,859 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,859 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,861 INFO [StoreOpener-f9a74e4ccdae036154df7dfc04494f2b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,864 INFO [StoreOpener-f9a74e4ccdae036154df7dfc04494f2b-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region f9a74e4ccdae036154df7dfc04494f2b columnFamilyName cf 2024-12-08T03:46:15,864 DEBUG [StoreOpener-f9a74e4ccdae036154df7dfc04494f2b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-08T03:46:15,865 INFO [StoreOpener-f9a74e4ccdae036154df7dfc04494f2b-1 {}] regionserver.HStore(327): Store=f9a74e4ccdae036154df7dfc04494f2b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-08T03:46:15,865 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,867 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,867 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,867 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,868 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,869 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,872 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-08T03:46:15,872 INFO [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened f9a74e4ccdae036154df7dfc04494f2b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=59567047, jitterRate=-0.11238183081150055}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-08T03:46:15,872 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:15,873 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for f9a74e4ccdae036154df7dfc04494f2b: Running coprocessor pre-open hook at 1733629575859Writing region info on filesystem at 1733629575859Initializing all the Stores at 1733629575860 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733629575861 (+1 ms)Cleaning up temporary data from old regions at 1733629575868 (+7 ms)Running coprocessor post-open hooks at 1733629575872 (+4 ms)Region opened successfully at 1733629575873 (+1 ms) 2024-12-08T03:46:15,874 INFO [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b., pid=6, masterSystemTime=1733629575849 2024-12-08T03:46:15,878 DEBUG [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:15,878 INFO [RS_OPEN_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:15,879 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=f9a74e4ccdae036154df7dfc04494f2b, regionState=OPEN, openSeqNum=2, regionLocation=32df4b4bb439,44085,1733629574352 2024-12-08T03:46:15,881 WARN [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37713 {}] assignment.AssignmentManager(1543): Unable to acquire lock for regionNode state=OPEN, location=32df4b4bb439,44085,1733629574352, table=TestHBaseWalOnEC, region=f9a74e4ccdae036154df7dfc04494f2b. It is likely that another thread is currently holding the lock. To avoid deadlock, skip execution for now. 2024-12-08T03:46:15,883 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure f9a74e4ccdae036154df7dfc04494f2b, server=32df4b4bb439,44085,1733629574352 because future has completed 2024-12-08T03:46:15,889 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-08T03:46:15,889 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure f9a74e4ccdae036154df7dfc04494f2b, server=32df4b4bb439,44085,1733629574352 in 191 msec 2024-12-08T03:46:15,894 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-08T03:46:15,894 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=f9a74e4ccdae036154df7dfc04494f2b, ASSIGN in 354 msec 2024-12-08T03:46:15,896 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-08T03:46:15,896 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733629575896"}]},"ts":"1733629575896"} 2024-12-08T03:46:15,900 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-08T03:46:15,902 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-08T03:46:15,905 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 413 msec 2024-12-08T03:46:16,123 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-08T03:46:16,124 INFO [RPCClient-NioEventLoopGroup-6-8 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-08T03:46:16,124 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-08T03:46:16,125 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-08T03:46:16,132 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-08T03:46:16,132 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-08T03:46:16,133 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-08T03:46:16,136 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b., hostname=32df4b4bb439,44085,1733629574352, seqNum=2] 2024-12-08T03:46:16,140 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-08T03:46:16,142 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-08T03:46:16,143 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-08T03:46:16,144 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-08T03:46:16,145 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-08T03:46:16,146 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-08T03:46:16,252 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-08T03:46:16,300 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=44085 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-08T03:46:16,300 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:16,300 INFO [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing f9a74e4ccdae036154df7dfc04494f2b 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-08T03:46:16,314 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b/.tmp/cf/61e43e69d31045f7bc24ef9a8a25c05a is 36, key is row/cf:cq/1733629576138/Put/seqid=0 2024-12-08T03:46:16,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741839_1015 (size=4787) 2024-12-08T03:46:16,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741839_1015 (size=4787) 2024-12-08T03:46:16,322 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741839_1015 (size=4787) 2024-12-08T03:46:16,322 INFO [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b/.tmp/cf/61e43e69d31045f7bc24ef9a8a25c05a 2024-12-08T03:46:16,330 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b/.tmp/cf/61e43e69d31045f7bc24ef9a8a25c05a as hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b/cf/61e43e69d31045f7bc24ef9a8a25c05a 2024-12-08T03:46:16,337 INFO [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b/cf/61e43e69d31045f7bc24ef9a8a25c05a, entries=1, sequenceid=5, filesize=4.7 K 2024-12-08T03:46:16,338 INFO [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for f9a74e4ccdae036154df7dfc04494f2b in 38ms, sequenceid=5, compaction requested=false 2024-12-08T03:46:16,339 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for f9a74e4ccdae036154df7dfc04494f2b: 2024-12-08T03:46:16,339 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:16,339 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32df4b4bb439:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-08T03:46:16,339 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-08T03:46:16,344 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-08T03:46:16,344 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 195 msec 2024-12-08T03:46:16,347 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 205 msec 2024-12-08T03:46:16,462 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37713 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-08T03:46:16,463 INFO [RPCClient-NioEventLoopGroup-6-8 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-08T03:46:16,467 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-08T03:46:16,468 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-08T03:46:16,468 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:16,468 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,468 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,468 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-08T03:46:16,468 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-08T03:46:16,468 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1625845178, stopped=false 2024-12-08T03:46:16,469 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=32df4b4bb439,37713,1733629574119 2024-12-08T03:46:16,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:16,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:16,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:16,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:16,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:16,515 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-08T03:46:16,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:16,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-08T03:46:16,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:16,516 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-08T03:46:16,516 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:16,516 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:16,516 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:16,516 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:16,516 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,516 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-08T03:46:16,516 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32df4b4bb439,43979,1733629574279' ***** 2024-12-08T03:46:16,516 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-08T03:46:16,516 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32df4b4bb439,34905,1733629574318' ***** 2024-12-08T03:46:16,516 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-08T03:46:16,517 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32df4b4bb439,44085,1733629574352' ***** 2024-12-08T03:46:16,517 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-08T03:46:16,517 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-08T03:46:16,517 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-08T03:46:16,517 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-08T03:46:16,517 INFO [RS:0;32df4b4bb439:43979 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-08T03:46:16,517 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-08T03:46:16,517 INFO [RS:0;32df4b4bb439:43979 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-08T03:46:16,517 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(959): stopping server 32df4b4bb439,43979,1733629574279 2024-12-08T03:46:16,517 INFO [RS:0;32df4b4bb439:43979 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-08T03:46:16,517 INFO [RS:1;32df4b4bb439:34905 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-08T03:46:16,517 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-08T03:46:16,517 INFO [RS:0;32df4b4bb439:43979 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;32df4b4bb439:43979. 2024-12-08T03:46:16,517 INFO [RS:1;32df4b4bb439:34905 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-08T03:46:16,517 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(959): stopping server 32df4b4bb439,34905,1733629574318 2024-12-08T03:46:16,517 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-08T03:46:16,517 DEBUG [RS:0;32df4b4bb439:43979 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:16,517 INFO [RS:1;32df4b4bb439:34905 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-08T03:46:16,517 DEBUG [RS:0;32df4b4bb439:43979 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,517 INFO [RS:1;32df4b4bb439:34905 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;32df4b4bb439:34905. 2024-12-08T03:46:16,517 DEBUG [RS:1;32df4b4bb439:34905 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:16,517 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(976): stopping server 32df4b4bb439,43979,1733629574279; all regions closed. 2024-12-08T03:46:16,517 INFO [RS:2;32df4b4bb439:44085 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-08T03:46:16,517 DEBUG [RS:1;32df4b4bb439:34905 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,518 INFO [RS:2;32df4b4bb439:44085 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-08T03:46:16,518 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(976): stopping server 32df4b4bb439,34905,1733629574318; all regions closed. 2024-12-08T03:46:16,518 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(3091): Received CLOSE for f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:16,518 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(959): stopping server 32df4b4bb439,44085,1733629574352 2024-12-08T03:46:16,518 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 INFO [RS:2;32df4b4bb439:44085 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-08T03:46:16,518 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 INFO [RS:2;32df4b4bb439:44085 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;32df4b4bb439:44085. 2024-12-08T03:46:16,518 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing f9a74e4ccdae036154df7dfc04494f2b, disabling compactions & flushes 2024-12-08T03:46:16,518 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 INFO [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:16,518 DEBUG [RS:2;32df4b4bb439:44085 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-08T03:46:16,518 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 DEBUG [RS:2;32df4b4bb439:44085 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,518 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-08T03:46:16,518 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-08T03:46:16,518 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,518 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:16,519 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-08T03:46:16,519 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. after waiting 0 ms 2024-12-08T03:46:16,519 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-08T03:46:16,519 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:16,519 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-08T03:46:16,519 DEBUG [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1325): Online Regions={f9a74e4ccdae036154df7dfc04494f2b=TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b., 1588230740=hbase:meta,,1.1588230740} 2024-12-08T03:46:16,519 DEBUG [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, f9a74e4ccdae036154df7dfc04494f2b 2024-12-08T03:46:16,519 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-08T03:46:16,519 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-08T03:46:16,519 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-08T03:46:16,519 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-08T03:46:16,519 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-08T03:46:16,520 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-08T03:46:16,521 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741834_1010 (size=93) 2024-12-08T03:46:16,521 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741834_1010 (size=93) 2024-12-08T03:46:16,521 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741834_1010 (size=93) 2024-12-08T03:46:16,524 DEBUG [RS:1;32df4b4bb439:34905 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/oldWALs 2024-12-08T03:46:16,524 INFO [RS:1;32df4b4bb439:34905 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 32df4b4bb439%2C34905%2C1733629574318:(num 1733629575027) 2024-12-08T03:46:16,524 DEBUG [RS:1;32df4b4bb439:34905 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,524 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:16,525 INFO [RS:1;32df4b4bb439:34905 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-08T03:46:16,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741835_1011 (size=93) 2024-12-08T03:46:16,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741835_1011 (size=93) 2024-12-08T03:46:16,526 INFO [RS:1;32df4b4bb439:34905 {}] hbase.ChoreService(370): Chore service for: regionserver/32df4b4bb439:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-08T03:46:16,526 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-08T03:46:16,526 INFO [regionserver/32df4b4bb439:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-08T03:46:16,526 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-08T03:46:16,526 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-08T03:46:16,526 INFO [RS:1;32df4b4bb439:34905 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-08T03:46:16,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741835_1011 (size=93) 2024-12-08T03:46:16,526 INFO [RS:1;32df4b4bb439:34905 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:34905 2024-12-08T03:46:16,528 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/default/TestHBaseWalOnEC/f9a74e4ccdae036154df7dfc04494f2b/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-08T03:46:16,529 DEBUG [RS:0;32df4b4bb439:43979 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/oldWALs 2024-12-08T03:46:16,529 INFO [RS:0;32df4b4bb439:43979 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 32df4b4bb439%2C43979%2C1733629574279:(num 1733629575027) 2024-12-08T03:46:16,529 INFO [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:16,529 DEBUG [RS:0;32df4b4bb439:43979 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,529 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:16,529 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for f9a74e4ccdae036154df7dfc04494f2b: Waiting for close lock at 1733629576518Running coprocessor pre-close hooks at 1733629576518Disabling compacts and flushes for region at 1733629576518Disabling writes for close at 1733629576519 (+1 ms)Writing region close event to WAL at 1733629576520 (+1 ms)Running coprocessor post-close hooks at 1733629576529 (+9 ms)Closed at 1733629576529 2024-12-08T03:46:16,529 INFO [RS:0;32df4b4bb439:43979 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-08T03:46:16,529 DEBUG [RS_CLOSE_REGION-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b. 2024-12-08T03:46:16,529 INFO [RS:0;32df4b4bb439:43979 {}] hbase.ChoreService(370): Chore service for: regionserver/32df4b4bb439:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-08T03:46:16,529 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-08T03:46:16,529 INFO [regionserver/32df4b4bb439:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-08T03:46:16,529 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-08T03:46:16,529 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-08T03:46:16,530 INFO [RS:0;32df4b4bb439:43979 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-08T03:46:16,530 INFO [RS:0;32df4b4bb439:43979 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:43979 2024-12-08T03:46:16,532 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32df4b4bb439,34905,1733629574318 2024-12-08T03:46:16,532 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-08T03:46:16,532 INFO [RS:1;32df4b4bb439:34905 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-08T03:46:16,538 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/info/ec8e6e5d150748c7b17d2b407817acc8 is 153, key is TestHBaseWalOnEC,,1733629575489.f9a74e4ccdae036154df7dfc04494f2b./info:regioninfo/1733629575879/Put/seqid=0 2024-12-08T03:46:16,540 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32df4b4bb439,43979,1733629574279 2024-12-08T03:46:16,540 INFO [RS:0;32df4b4bb439:43979 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-08T03:46:16,540 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32df4b4bb439,43979,1733629574279] 2024-12-08T03:46:16,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741840_1016 (size=6637) 2024-12-08T03:46:16,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741840_1016 (size=6637) 2024-12-08T03:46:16,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741840_1016 (size=6637) 2024-12-08T03:46:16,545 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/info/ec8e6e5d150748c7b17d2b407817acc8 2024-12-08T03:46:16,557 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32df4b4bb439,43979,1733629574279 already deleted, retry=false 2024-12-08T03:46:16,557 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32df4b4bb439,43979,1733629574279 expired; onlineServers=2 2024-12-08T03:46:16,557 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32df4b4bb439,34905,1733629574318] 2024-12-08T03:46:16,565 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32df4b4bb439,34905,1733629574318 already deleted, retry=false 2024-12-08T03:46:16,565 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32df4b4bb439,34905,1733629574318 expired; onlineServers=1 2024-12-08T03:46:16,572 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/ns/6ec39030f6d1479483f1af2528f817a4 is 43, key is default/ns:d/1733629575445/Put/seqid=0 2024-12-08T03:46:16,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741841_1017 (size=5153) 2024-12-08T03:46:16,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741841_1017 (size=5153) 2024-12-08T03:46:16,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741841_1017 (size=5153) 2024-12-08T03:46:16,580 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/ns/6ec39030f6d1479483f1af2528f817a4 2024-12-08T03:46:16,589 INFO [regionserver/32df4b4bb439:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:16,592 INFO [regionserver/32df4b4bb439:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:16,594 INFO [regionserver/32df4b4bb439:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:16,600 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/table/1fe00c013bcd4f3790868be5d4e77e9d is 52, key is TestHBaseWalOnEC/table:state/1733629575896/Put/seqid=0 2024-12-08T03:46:16,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741842_1018 (size=5249) 2024-12-08T03:46:16,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741842_1018 (size=5249) 2024-12-08T03:46:16,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741842_1018 (size=5249) 2024-12-08T03:46:16,607 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/table/1fe00c013bcd4f3790868be5d4e77e9d 2024-12-08T03:46:16,615 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/info/ec8e6e5d150748c7b17d2b407817acc8 as hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/info/ec8e6e5d150748c7b17d2b407817acc8 2024-12-08T03:46:16,623 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/info/ec8e6e5d150748c7b17d2b407817acc8, entries=10, sequenceid=11, filesize=6.5 K 2024-12-08T03:46:16,625 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/ns/6ec39030f6d1479483f1af2528f817a4 as hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/ns/6ec39030f6d1479483f1af2528f817a4 2024-12-08T03:46:16,633 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/ns/6ec39030f6d1479483f1af2528f817a4, entries=2, sequenceid=11, filesize=5.0 K 2024-12-08T03:46:16,635 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/.tmp/table/1fe00c013bcd4f3790868be5d4e77e9d as hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/table/1fe00c013bcd4f3790868be5d4e77e9d 2024-12-08T03:46:16,644 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/table/1fe00c013bcd4f3790868be5d4e77e9d, entries=2, sequenceid=11, filesize=5.1 K 2024-12-08T03:46:16,646 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 126ms, sequenceid=11, compaction requested=false 2024-12-08T03:46:16,649 INFO [RS:1;32df4b4bb439:34905 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-08T03:46:16,649 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:16,649 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34905-0x1000395ba6f0002, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:16,649 INFO [RS:1;32df4b4bb439:34905 {}] regionserver.HRegionServer(1031): Exiting; stopping=32df4b4bb439,34905,1733629574318; zookeeper connection closed. 2024-12-08T03:46:16,649 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@6762f2a2 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@6762f2a2 2024-12-08T03:46:16,652 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-08T03:46:16,653 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-08T03:46:16,653 INFO [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-08T03:46:16,653 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733629576519Running coprocessor pre-close hooks at 1733629576519Disabling compacts and flushes for region at 1733629576519Disabling writes for close at 1733629576519Obtaining lock to block concurrent updates at 1733629576520 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1733629576520Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733629576520Flushing stores of hbase:meta,,1.1588230740 at 1733629576521 (+1 ms)Flushing 1588230740/info: creating writer at 1733629576521Flushing 1588230740/info: appending metadata at 1733629576537 (+16 ms)Flushing 1588230740/info: closing flushed file at 1733629576537Flushing 1588230740/ns: creating writer at 1733629576553 (+16 ms)Flushing 1588230740/ns: appending metadata at 1733629576572 (+19 ms)Flushing 1588230740/ns: closing flushed file at 1733629576572Flushing 1588230740/table: creating writer at 1733629576586 (+14 ms)Flushing 1588230740/table: appending metadata at 1733629576599 (+13 ms)Flushing 1588230740/table: closing flushed file at 1733629576599Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@21be1dbd: reopening flushed file at 1733629576614 (+15 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@68637e47: reopening flushed file at 1733629576623 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7bc39052: reopening flushed file at 1733629576634 (+11 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 126ms, sequenceid=11, compaction requested=false at 1733629576646 (+12 ms)Writing region close event to WAL at 1733629576647 (+1 ms)Running coprocessor post-close hooks at 1733629576652 (+5 ms)Closed at 1733629576653 (+1 ms) 2024-12-08T03:46:16,653 DEBUG [RS_CLOSE_META-regionserver/32df4b4bb439:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-08T03:46:16,657 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:16,657 INFO [RS:0;32df4b4bb439:43979 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-08T03:46:16,657 INFO [RS:0;32df4b4bb439:43979 {}] regionserver.HRegionServer(1031): Exiting; stopping=32df4b4bb439,43979,1733629574279; zookeeper connection closed. 2024-12-08T03:46:16,657 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43979-0x1000395ba6f0001, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:16,657 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@677281a7 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@677281a7 2024-12-08T03:46:16,719 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(976): stopping server 32df4b4bb439,44085,1733629574352; all regions closed. 2024-12-08T03:46:16,720 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,720 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,720 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,720 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,721 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741836_1012 (size=2751) 2024-12-08T03:46:16,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741836_1012 (size=2751) 2024-12-08T03:46:16,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741836_1012 (size=2751) 2024-12-08T03:46:16,726 DEBUG [RS:2;32df4b4bb439:44085 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/oldWALs 2024-12-08T03:46:16,726 INFO [RS:2;32df4b4bb439:44085 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 32df4b4bb439%2C44085%2C1733629574352.meta:.meta(num 1733629575372) 2024-12-08T03:46:16,726 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,726 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,727 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,727 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,727 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741833_1009 (size=1298) 2024-12-08T03:46:16,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741833_1009 (size=1298) 2024-12-08T03:46:16,730 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741833_1009 (size=1298) 2024-12-08T03:46:16,732 DEBUG [RS:2;32df4b4bb439:44085 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/oldWALs 2024-12-08T03:46:16,732 INFO [RS:2;32df4b4bb439:44085 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 32df4b4bb439%2C44085%2C1733629574352:(num 1733629575025) 2024-12-08T03:46:16,732 DEBUG [RS:2;32df4b4bb439:44085 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-08T03:46:16,732 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.LeaseManager(133): Closed leases 2024-12-08T03:46:16,732 INFO [RS:2;32df4b4bb439:44085 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-08T03:46:16,732 INFO [RS:2;32df4b4bb439:44085 {}] hbase.ChoreService(370): Chore service for: regionserver/32df4b4bb439:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-08T03:46:16,732 INFO [RS:2;32df4b4bb439:44085 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-08T03:46:16,733 INFO [regionserver/32df4b4bb439:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-08T03:46:16,733 INFO [RS:2;32df4b4bb439:44085 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:44085 2024-12-08T03:46:16,757 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32df4b4bb439,44085,1733629574352 2024-12-08T03:46:16,757 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-08T03:46:16,757 INFO [RS:2;32df4b4bb439:44085 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-08T03:46:16,765 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32df4b4bb439,44085,1733629574352] 2024-12-08T03:46:16,773 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32df4b4bb439,44085,1733629574352 already deleted, retry=false 2024-12-08T03:46:16,773 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32df4b4bb439,44085,1733629574352 expired; onlineServers=0 2024-12-08T03:46:16,774 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '32df4b4bb439,37713,1733629574119' ***** 2024-12-08T03:46:16,774 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-08T03:46:16,774 INFO [M:0;32df4b4bb439:37713 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-08T03:46:16,774 INFO [M:0;32df4b4bb439:37713 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-08T03:46:16,774 DEBUG [M:0;32df4b4bb439:37713 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-08T03:46:16,774 DEBUG [M:0;32df4b4bb439:37713 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-08T03:46:16,774 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-08T03:46:16,774 DEBUG [master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.small.0-1733629574702 {}] cleaner.HFileCleaner(306): Exit Thread[master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.small.0-1733629574702,5,FailOnTimeoutGroup] 2024-12-08T03:46:16,774 DEBUG [master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.large.0-1733629574701 {}] cleaner.HFileCleaner(306): Exit Thread[master/32df4b4bb439:0:becomeActiveMaster-HFileCleaner.large.0-1733629574701,5,FailOnTimeoutGroup] 2024-12-08T03:46:16,774 INFO [M:0;32df4b4bb439:37713 {}] hbase.ChoreService(370): Chore service for: master/32df4b4bb439:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-08T03:46:16,774 INFO [M:0;32df4b4bb439:37713 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-08T03:46:16,774 DEBUG [M:0;32df4b4bb439:37713 {}] master.HMaster(1795): Stopping service threads 2024-12-08T03:46:16,774 INFO [M:0;32df4b4bb439:37713 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-08T03:46:16,775 INFO [M:0;32df4b4bb439:37713 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-08T03:46:16,775 INFO [M:0;32df4b4bb439:37713 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-08T03:46:16,775 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-08T03:46:16,782 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-08T03:46:16,782 DEBUG [M:0;32df4b4bb439:37713 {}] zookeeper.ZKUtil(347): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-08T03:46:16,782 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-08T03:46:16,782 WARN [M:0;32df4b4bb439:37713 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-08T03:46:16,783 INFO [M:0;32df4b4bb439:37713 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/.lastflushedseqids 2024-12-08T03:46:16,785 WARN [IPC Server handler 2 on default port 41623 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-08T03:46:16,785 WARN [IPC Server handler 2 on default port 41623 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-08T03:46:16,785 WARN [IPC Server handler 2 on default port 41623 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-08T03:46:16,794 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741843_1019 (size=127) 2024-12-08T03:46:16,794 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741843_1019 (size=127) 2024-12-08T03:46:16,794 INFO [M:0;32df4b4bb439:37713 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-08T03:46:16,795 INFO [M:0;32df4b4bb439:37713 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-08T03:46:16,795 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-08T03:46:16,795 INFO [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:16,795 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:16,795 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-08T03:46:16,795 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:16,795 INFO [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.83 KB heapSize=34.12 KB 2024-12-08T03:46:16,817 DEBUG [M:0;32df4b4bb439:37713 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/2812b031bc7546dbbb08e8affac4b994 is 82, key is hbase:meta,,1/info:regioninfo/1733629575408/Put/seqid=0 2024-12-08T03:46:16,819 WARN [IPC Server handler 3 on default port 41623 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-08T03:46:16,819 WARN [IPC Server handler 3 on default port 41623 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-08T03:46:16,819 WARN [IPC Server handler 3 on default port 41623 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-08T03:46:16,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741844_1020 (size=5672) 2024-12-08T03:46:16,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741844_1020 (size=5672) 2024-12-08T03:46:16,825 INFO [M:0;32df4b4bb439:37713 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/2812b031bc7546dbbb08e8affac4b994 2024-12-08T03:46:16,845 DEBUG [M:0;32df4b4bb439:37713 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a7fbfaab90d04a63b29c4c352607a146 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733629575904/Put/seqid=0 2024-12-08T03:46:16,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741845_1021 (size=6439) 2024-12-08T03:46:16,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741845_1021 (size=6439) 2024-12-08T03:46:16,858 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741845_1021 (size=6439) 2024-12-08T03:46:16,860 INFO [M:0;32df4b4bb439:37713 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a7fbfaab90d04a63b29c4c352607a146 2024-12-08T03:46:16,865 INFO [RS:2;32df4b4bb439:44085 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-08T03:46:16,865 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:16,865 INFO [RS:2;32df4b4bb439:44085 {}] regionserver.HRegionServer(1031): Exiting; stopping=32df4b4bb439,44085,1733629574352; zookeeper connection closed. 2024-12-08T03:46:16,865 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44085-0x1000395ba6f0003, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:16,866 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@2760972c {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@2760972c 2024-12-08T03:46:16,866 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-08T03:46:16,881 DEBUG [M:0;32df4b4bb439:37713 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ee0defcb3f9f4ed19a96ed97ee68cbec is 69, key is 32df4b4bb439,34905,1733629574318/rs:state/1733629574830/Put/seqid=0 2024-12-08T03:46:16,882 WARN [IPC Server handler 2 on default port 41623 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-08T03:46:16,882 WARN [IPC Server handler 2 on default port 41623 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-08T03:46:16,882 WARN [IPC Server handler 2 on default port 41623 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-08T03:46:16,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741846_1022 (size=5294) 2024-12-08T03:46:16,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741846_1022 (size=5294) 2024-12-08T03:46:16,887 INFO [M:0;32df4b4bb439:37713 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ee0defcb3f9f4ed19a96ed97ee68cbec 2024-12-08T03:46:16,894 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/2812b031bc7546dbbb08e8affac4b994 as hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/2812b031bc7546dbbb08e8affac4b994 2024-12-08T03:46:16,900 INFO [M:0;32df4b4bb439:37713 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/2812b031bc7546dbbb08e8affac4b994, entries=8, sequenceid=72, filesize=5.5 K 2024-12-08T03:46:16,902 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a7fbfaab90d04a63b29c4c352607a146 as hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/a7fbfaab90d04a63b29c4c352607a146 2024-12-08T03:46:16,907 INFO [M:0;32df4b4bb439:37713 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/a7fbfaab90d04a63b29c4c352607a146, entries=8, sequenceid=72, filesize=6.3 K 2024-12-08T03:46:16,908 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ee0defcb3f9f4ed19a96ed97ee68cbec as hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/ee0defcb3f9f4ed19a96ed97ee68cbec 2024-12-08T03:46:16,915 INFO [M:0;32df4b4bb439:37713 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41623/user/jenkins/test-data/f128385e-8250-54cd-4f3f-3255487075fe/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/ee0defcb3f9f4ed19a96ed97ee68cbec, entries=3, sequenceid=72, filesize=5.2 K 2024-12-08T03:46:16,916 INFO [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 121ms, sequenceid=72, compaction requested=false 2024-12-08T03:46:16,917 INFO [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-08T03:46:16,917 DEBUG [M:0;32df4b4bb439:37713 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733629576795Disabling compacts and flushes for region at 1733629576795Disabling writes for close at 1733629576795Obtaining lock to block concurrent updates at 1733629576795Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733629576795Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27471, getHeapSize=34872, getOffHeapSize=0, getCellsCount=85 at 1733629576796 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733629576797 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733629576797Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733629576817 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733629576817Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733629576831 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733629576845 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733629576845Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733629576866 (+21 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733629576881 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733629576881Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@46cb0bb0: reopening flushed file at 1733629576893 (+12 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@142532d9: reopening flushed file at 1733629576901 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@b3bec10: reopening flushed file at 1733629576907 (+6 ms)Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 121ms, sequenceid=72, compaction requested=false at 1733629576916 (+9 ms)Writing region close event to WAL at 1733629576917 (+1 ms)Closed at 1733629576917 2024-12-08T03:46:16,918 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,918 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,918 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,918 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,918 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-08T03:46:16,920 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32999 is added to blk_1073741830_1006 (size=32674) 2024-12-08T03:46:16,920 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37631 is added to blk_1073741830_1006 (size=32674) 2024-12-08T03:46:16,921 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32941 is added to blk_1073741830_1006 (size=32674) 2024-12-08T03:46:16,922 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-08T03:46:16,922 INFO [M:0;32df4b4bb439:37713 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-08T03:46:16,922 INFO [M:0;32df4b4bb439:37713 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:37713 2024-12-08T03:46:16,922 INFO [M:0;32df4b4bb439:37713 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-08T03:46:17,041 INFO [M:0;32df4b4bb439:37713 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-08T03:46:17,041 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:17,041 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37713-0x1000395ba6f0000, quorum=127.0.0.1:53394, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-08T03:46:17,044 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@611d8bdf{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:17,045 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@8e73203{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-08T03:46:17,045 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-08T03:46:17,045 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@659fda5a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-08T03:46:17,046 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@69168ef4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir/,STOPPED} 2024-12-08T03:46:17,048 WARN [BP-1091900809-172.17.0.2-1733629572156 heartbeating to localhost/127.0.0.1:41623 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-08T03:46:17,048 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-08T03:46:17,048 WARN [BP-1091900809-172.17.0.2-1733629572156 heartbeating to localhost/127.0.0.1:41623 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1091900809-172.17.0.2-1733629572156 (Datanode Uuid a9b549fc-4f7d-4092-9b79-c1d7410b62b2) service to localhost/127.0.0.1:41623 2024-12-08T03:46:17,048 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-08T03:46:17,049 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data5/current/BP-1091900809-172.17.0.2-1733629572156 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:17,050 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data6/current/BP-1091900809-172.17.0.2-1733629572156 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:17,050 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-08T03:46:17,053 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1c5541de{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:17,053 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@624e70a4{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-08T03:46:17,053 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-08T03:46:17,054 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@143655f2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-08T03:46:17,054 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7ddf5a9c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir/,STOPPED} 2024-12-08T03:46:17,055 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-08T03:46:17,055 WARN [BP-1091900809-172.17.0.2-1733629572156 heartbeating to localhost/127.0.0.1:41623 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-08T03:46:17,055 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-08T03:46:17,055 WARN [BP-1091900809-172.17.0.2-1733629572156 heartbeating to localhost/127.0.0.1:41623 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1091900809-172.17.0.2-1733629572156 (Datanode Uuid 298bfc20-1fd4-4c55-b99d-b422c5e9632f) service to localhost/127.0.0.1:41623 2024-12-08T03:46:17,056 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data3/current/BP-1091900809-172.17.0.2-1733629572156 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:17,056 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data4/current/BP-1091900809-172.17.0.2-1733629572156 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:17,056 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-08T03:46:17,059 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@12529ca{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-08T03:46:17,059 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@847d000{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-08T03:46:17,059 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-08T03:46:17,059 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3ae83c3f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-08T03:46:17,059 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2aa8c76d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir/,STOPPED} 2024-12-08T03:46:17,061 WARN [BP-1091900809-172.17.0.2-1733629572156 heartbeating to localhost/127.0.0.1:41623 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-08T03:46:17,061 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-08T03:46:17,061 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-08T03:46:17,061 WARN [BP-1091900809-172.17.0.2-1733629572156 heartbeating to localhost/127.0.0.1:41623 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1091900809-172.17.0.2-1733629572156 (Datanode Uuid 322004e8-176c-4f0a-899b-eaab4d57e321) service to localhost/127.0.0.1:41623 2024-12-08T03:46:17,061 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data1/current/BP-1091900809-172.17.0.2-1733629572156 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:17,061 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/cluster_476a8bb7-9f25-d788-cd01-185ec2fd807d/data/data2/current/BP-1091900809-172.17.0.2-1733629572156 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-08T03:46:17,062 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-08T03:46:17,067 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4f148449{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-08T03:46:17,067 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4abfe850{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-08T03:46:17,067 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-08T03:46:17,068 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3950f25b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-08T03:46:17,068 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@377873f6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/7f911b7a-81cc-89a7-3dfe-f6c92118091a/hadoop.log.dir/,STOPPED} 2024-12-08T03:46:17,074 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-08T03:46:17,100 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-08T03:46:17,106 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=147 (was 90) - Thread LEAK? -, OpenFileDescriptor=518 (was 439) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=170 (was 141) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=17325 (was 17505)