2024-12-04 09:42:59,219 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-04 09:42:59,229 main DEBUG Took 0.008659 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-04 09:42:59,230 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-04 09:42:59,230 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-04 09:42:59,231 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-04 09:42:59,232 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,241 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-04 09:42:59,257 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,259 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,260 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,260 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,261 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,261 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,263 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,263 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,264 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,264 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,265 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,266 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,266 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,267 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,267 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,268 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,268 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,269 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,269 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,269 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,270 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,270 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,271 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,271 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-04 09:42:59,272 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,272 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-04 09:42:59,274 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-04 09:42:59,275 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-04 09:42:59,277 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-04 09:42:59,277 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-04 09:42:59,278 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-04 09:42:59,279 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-04 09:42:59,286 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-04 09:42:59,289 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-04 09:42:59,291 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-04 09:42:59,291 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-04 09:42:59,292 main DEBUG createAppenders(={Console}) 2024-12-04 09:42:59,293 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-04 09:42:59,293 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-04 09:42:59,293 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-04 09:42:59,294 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-04 09:42:59,294 main DEBUG OutputStream closed 2024-12-04 09:42:59,294 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-04 09:42:59,294 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-04 09:42:59,295 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-04 09:42:59,361 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-04 09:42:59,363 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-04 09:42:59,365 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-04 09:42:59,366 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-04 09:42:59,367 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-04 09:42:59,368 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-04 09:42:59,368 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-04 09:42:59,369 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-04 09:42:59,370 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-04 09:42:59,370 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-04 09:42:59,370 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-04 09:42:59,371 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-04 09:42:59,371 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-04 09:42:59,372 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-04 09:42:59,372 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-04 09:42:59,372 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-04 09:42:59,373 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-04 09:42:59,373 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-04 09:42:59,376 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-04 09:42:59,376 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-04 09:42:59,376 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-04 09:42:59,377 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-04T09:42:59,392 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-04 09:42:59,394 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-04 09:42:59,395 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-04T09:42:59,647 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752 2024-12-04T09:42:59,669 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79, deleteOnExit=true 2024-12-04T09:42:59,670 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/test.cache.data in system properties and HBase conf 2024-12-04T09:42:59,671 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.tmp.dir in system properties and HBase conf 2024-12-04T09:42:59,671 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir in system properties and HBase conf 2024-12-04T09:42:59,672 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-04T09:42:59,672 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-04T09:42:59,673 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-04T09:42:59,749 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-04T09:42:59,824 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-04T09:42:59,827 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-04T09:42:59,827 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-04T09:42:59,828 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-04T09:42:59,828 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-04T09:42:59,828 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-04T09:42:59,829 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-04T09:42:59,829 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-04T09:42:59,829 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-04T09:42:59,830 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-04T09:42:59,830 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/nfs.dump.dir in system properties and HBase conf 2024-12-04T09:42:59,830 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/java.io.tmpdir in system properties and HBase conf 2024-12-04T09:42:59,831 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-04T09:42:59,831 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-04T09:42:59,831 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-04T09:43:00,665 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-04T09:43:00,728 INFO [Time-limited test {}] log.Log(170): Logging initialized @2091ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-04T09:43:00,785 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:00,836 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-04T09:43:00,853 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-04T09:43:00,853 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-04T09:43:00,854 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-04T09:43:00,866 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:00,868 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir/,AVAILABLE} 2024-12-04T09:43:00,869 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-04T09:43:01,025 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@76e4c45c{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/java.io.tmpdir/jetty-localhost-41455-hadoop-hdfs-3_4_1-tests_jar-_-any-16528508487487204490/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-04T09:43:01,031 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:41455} 2024-12-04T09:43:01,031 INFO [Time-limited test {}] server.Server(415): Started @2394ms 2024-12-04T09:43:01,497 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:01,503 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-04T09:43:01,504 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-04T09:43:01,504 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-04T09:43:01,504 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-04T09:43:01,505 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@550154bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir/,AVAILABLE} 2024-12-04T09:43:01,506 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1a2478ad{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-04T09:43:01,597 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4839957b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/java.io.tmpdir/jetty-localhost-42125-hadoop-hdfs-3_4_1-tests_jar-_-any-2266371838066842398/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:01,597 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5306f615{HTTP/1.1, (http/1.1)}{localhost:42125} 2024-12-04T09:43:01,598 INFO [Time-limited test {}] server.Server(415): Started @2961ms 2024-12-04T09:43:01,641 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-04T09:43:01,732 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:01,736 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-04T09:43:01,743 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-04T09:43:01,743 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-04T09:43:01,743 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-04T09:43:01,746 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6463ad04{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir/,AVAILABLE} 2024-12-04T09:43:01,747 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7fa8fa5c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-04T09:43:01,849 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1c6b8f01{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/java.io.tmpdir/jetty-localhost-46369-hadoop-hdfs-3_4_1-tests_jar-_-any-11022748167176458230/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:01,850 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11f28dd2{HTTP/1.1, (http/1.1)}{localhost:46369} 2024-12-04T09:43:01,850 INFO [Time-limited test {}] server.Server(415): Started @3213ms 2024-12-04T09:43:01,852 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-04T09:43:01,883 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:01,888 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-04T09:43:01,890 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-04T09:43:01,890 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-04T09:43:01,891 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-04T09:43:01,892 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@c62369b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir/,AVAILABLE} 2024-12-04T09:43:01,892 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@24f92c39{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-04T09:43:01,990 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2e59159d{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/java.io.tmpdir/jetty-localhost-40725-hadoop-hdfs-3_4_1-tests_jar-_-any-698435857371630880/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:01,991 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@a8e922f{HTTP/1.1, (http/1.1)}{localhost:40725} 2024-12-04T09:43:01,991 INFO [Time-limited test {}] server.Server(415): Started @3355ms 2024-12-04T09:43:01,993 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-04T09:43:02,694 WARN [Thread-123 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data4/current/BP-1422175098-172.17.0.2-1733305380290/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:02,694 WARN [Thread-124 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data1/current/BP-1422175098-172.17.0.2-1733305380290/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:02,694 WARN [Thread-125 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data2/current/BP-1422175098-172.17.0.2-1733305380290/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:02,694 WARN [Thread-122 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data3/current/BP-1422175098-172.17.0.2-1733305380290/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:02,730 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-04T09:43:02,730 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-04T09:43:02,767 WARN [Thread-143 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data6/current/BP-1422175098-172.17.0.2-1733305380290/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:02,767 WARN [Thread-142 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data5/current/BP-1422175098-172.17.0.2-1733305380290/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:02,777 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x104c8c23e8e88a81 with lease ID 0x43cfccec0fb038fd: Processing first storage report for DS-57987938-e5b7-4626-9be2-fc660a712f31 from datanode DatanodeRegistration(127.0.0.1:34441, datanodeUuid=1566c7e3-5029-4d80-98d9-753760a78791, infoPort=37275, infoSecurePort=0, ipcPort=46801, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290) 2024-12-04T09:43:02,778 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x104c8c23e8e88a81 with lease ID 0x43cfccec0fb038fd: from storage DS-57987938-e5b7-4626-9be2-fc660a712f31 node DatanodeRegistration(127.0.0.1:34441, datanodeUuid=1566c7e3-5029-4d80-98d9-753760a78791, infoPort=37275, infoSecurePort=0, ipcPort=46801, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-04T09:43:02,778 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc67d813c2b708a0d with lease ID 0x43cfccec0fb038fe: Processing first storage report for DS-1df1c0e0-adcb-42a3-8d8b-7e6526c2c617 from datanode DatanodeRegistration(127.0.0.1:42541, datanodeUuid=5f745f9f-5b17-457a-b10d-ea3589192b67, infoPort=35933, infoSecurePort=0, ipcPort=34815, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290) 2024-12-04T09:43:02,778 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc67d813c2b708a0d with lease ID 0x43cfccec0fb038fe: from storage DS-1df1c0e0-adcb-42a3-8d8b-7e6526c2c617 node DatanodeRegistration(127.0.0.1:42541, datanodeUuid=5f745f9f-5b17-457a-b10d-ea3589192b67, infoPort=35933, infoSecurePort=0, ipcPort=34815, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:02,778 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x104c8c23e8e88a81 with lease ID 0x43cfccec0fb038fd: Processing first storage report for DS-8234dd40-75ce-4122-b75d-994f4c181ebc from datanode DatanodeRegistration(127.0.0.1:34441, datanodeUuid=1566c7e3-5029-4d80-98d9-753760a78791, infoPort=37275, infoSecurePort=0, ipcPort=46801, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290) 2024-12-04T09:43:02,779 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x104c8c23e8e88a81 with lease ID 0x43cfccec0fb038fd: from storage DS-8234dd40-75ce-4122-b75d-994f4c181ebc node DatanodeRegistration(127.0.0.1:34441, datanodeUuid=1566c7e3-5029-4d80-98d9-753760a78791, infoPort=37275, infoSecurePort=0, ipcPort=46801, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:02,779 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc67d813c2b708a0d with lease ID 0x43cfccec0fb038fe: Processing first storage report for DS-74668357-f8d4-4077-9371-171df5a224f3 from datanode DatanodeRegistration(127.0.0.1:42541, datanodeUuid=5f745f9f-5b17-457a-b10d-ea3589192b67, infoPort=35933, infoSecurePort=0, ipcPort=34815, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290) 2024-12-04T09:43:02,779 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc67d813c2b708a0d with lease ID 0x43cfccec0fb038fe: from storage DS-74668357-f8d4-4077-9371-171df5a224f3 node DatanodeRegistration(127.0.0.1:42541, datanodeUuid=5f745f9f-5b17-457a-b10d-ea3589192b67, infoPort=35933, infoSecurePort=0, ipcPort=34815, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:02,787 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-04T09:43:02,791 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2ce8f3ba8b8a763 with lease ID 0x43cfccec0fb038ff: Processing first storage report for DS-2e03073c-8b4e-4ef7-85e4-776e29ea76bb from datanode DatanodeRegistration(127.0.0.1:42163, datanodeUuid=6df7b36d-cc7c-4d72-b69d-8dd2ee15e86d, infoPort=42869, infoSecurePort=0, ipcPort=41645, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290) 2024-12-04T09:43:02,791 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2ce8f3ba8b8a763 with lease ID 0x43cfccec0fb038ff: from storage DS-2e03073c-8b4e-4ef7-85e4-776e29ea76bb node DatanodeRegistration(127.0.0.1:42163, datanodeUuid=6df7b36d-cc7c-4d72-b69d-8dd2ee15e86d, infoPort=42869, infoSecurePort=0, ipcPort=41645, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-04T09:43:02,792 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2ce8f3ba8b8a763 with lease ID 0x43cfccec0fb038ff: Processing first storage report for DS-af9f432d-f0fa-4b52-859f-a7d5c5e7fd90 from datanode DatanodeRegistration(127.0.0.1:42163, datanodeUuid=6df7b36d-cc7c-4d72-b69d-8dd2ee15e86d, infoPort=42869, infoSecurePort=0, ipcPort=41645, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290) 2024-12-04T09:43:02,792 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2ce8f3ba8b8a763 with lease ID 0x43cfccec0fb038ff: from storage DS-af9f432d-f0fa-4b52-859f-a7d5c5e7fd90 node DatanodeRegistration(127.0.0.1:42163, datanodeUuid=6df7b36d-cc7c-4d72-b69d-8dd2ee15e86d, infoPort=42869, infoSecurePort=0, ipcPort=41645, storageInfo=lv=-57;cid=testClusterID;nsid=1245556135;c=1733305380290), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:02,872 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752 2024-12-04T09:43:02,935 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-04T09:43:02,981 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=161, OpenFileDescriptor=393, MaxFileDescriptor=1048576, SystemLoadAverage=326, ProcessCount=11, AvailableMemoryMB=10705 2024-12-04T09:43:02,982 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-04T09:43:02,988 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-04T09:43:03,052 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/zookeeper_0, clientPort=60962, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-04T09:43:03,061 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=60962 2024-12-04T09:43:03,077 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:03,080 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:03,152 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:03,153 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:03,196 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:53166 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53166 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:03,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-04T09:43:03,616 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:03,629 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241 with version=8 2024-12-04T09:43:03,629 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/hbase-staging 2024-12-04T09:43:03,708 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-04T09:43:03,925 INFO [Time-limited test {}] client.ConnectionUtils(128): master/84486a41f81c:0 server-side Connection retries=45 2024-12-04T09:43:03,934 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:03,934 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:03,939 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-04T09:43:03,939 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:03,939 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-04T09:43:04,087 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-04T09:43:04,142 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-04T09:43:04,151 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-04T09:43:04,154 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-04T09:43:04,175 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 64218 (auto-detected) 2024-12-04T09:43:04,176 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-04T09:43:04,191 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:34373 2024-12-04T09:43:04,209 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:34373 connecting to ZooKeeper ensemble=127.0.0.1:60962 2024-12-04T09:43:04,310 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:343730x0, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-04T09:43:04,313 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:34373-0x101a102daf10000 connected 2024-12-04T09:43:04,391 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,395 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,405 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:04,409 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241, hbase.cluster.distributed=false 2024-12-04T09:43:04,429 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-04T09:43:04,434 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34373 2024-12-04T09:43:04,434 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34373 2024-12-04T09:43:04,435 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34373 2024-12-04T09:43:04,435 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34373 2024-12-04T09:43:04,435 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34373 2024-12-04T09:43:04,519 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/84486a41f81c:0 server-side Connection retries=45 2024-12-04T09:43:04,520 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,521 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,521 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-04T09:43:04,521 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,521 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-04T09:43:04,523 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-04T09:43:04,525 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-04T09:43:04,526 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:40823 2024-12-04T09:43:04,527 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:40823 connecting to ZooKeeper ensemble=127.0.0.1:60962 2024-12-04T09:43:04,528 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,530 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:408230x0, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-04T09:43:04,540 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:04,540 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:40823-0x101a102daf10001 connected 2024-12-04T09:43:04,544 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-04T09:43:04,552 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-04T09:43:04,555 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-04T09:43:04,559 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-04T09:43:04,560 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40823 2024-12-04T09:43:04,560 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40823 2024-12-04T09:43:04,563 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40823 2024-12-04T09:43:04,564 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40823 2024-12-04T09:43:04,564 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40823 2024-12-04T09:43:04,579 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/84486a41f81c:0 server-side Connection retries=45 2024-12-04T09:43:04,579 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,580 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,580 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-04T09:43:04,580 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,581 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-04T09:43:04,581 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-04T09:43:04,581 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-04T09:43:04,582 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:34771 2024-12-04T09:43:04,584 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:34771 connecting to ZooKeeper ensemble=127.0.0.1:60962 2024-12-04T09:43:04,586 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,589 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,604 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:347710x0, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-04T09:43:04,605 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:34771-0x101a102daf10002 connected 2024-12-04T09:43:04,605 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:04,605 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-04T09:43:04,606 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-04T09:43:04,607 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-04T09:43:04,610 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-04T09:43:04,610 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34771 2024-12-04T09:43:04,611 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34771 2024-12-04T09:43:04,611 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34771 2024-12-04T09:43:04,612 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34771 2024-12-04T09:43:04,612 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34771 2024-12-04T09:43:04,626 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/84486a41f81c:0 server-side Connection retries=45 2024-12-04T09:43:04,626 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,626 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,627 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-04T09:43:04,627 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:04,627 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-04T09:43:04,627 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-04T09:43:04,627 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-04T09:43:04,628 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39799 2024-12-04T09:43:04,630 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39799 connecting to ZooKeeper ensemble=127.0.0.1:60962 2024-12-04T09:43:04,632 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,634 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,645 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:397990x0, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-04T09:43:04,646 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:397990x0, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:04,646 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39799-0x101a102daf10003 connected 2024-12-04T09:43:04,647 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-04T09:43:04,648 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-04T09:43:04,649 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-04T09:43:04,650 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-04T09:43:04,652 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39799 2024-12-04T09:43:04,652 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39799 2024-12-04T09:43:04,653 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39799 2024-12-04T09:43:04,654 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39799 2024-12-04T09:43:04,654 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39799 2024-12-04T09:43:04,668 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;84486a41f81c:34373 2024-12-04T09:43:04,668 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/84486a41f81c,34373,1733305383783 2024-12-04T09:43:04,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:04,682 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:04,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:04,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:04,684 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/84486a41f81c,34373,1733305383783 2024-12-04T09:43:04,720 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-04T09:43:04,720 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:04,720 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-04T09:43:04,720 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-04T09:43:04,720 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:04,720 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:04,720 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:04,721 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-04T09:43:04,722 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/84486a41f81c,34373,1733305383783 from backup master directory 2024-12-04T09:43:04,731 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:04,731 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:04,731 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/84486a41f81c,34373,1733305383783 2024-12-04T09:43:04,731 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:04,732 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:04,732 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-04T09:43:04,733 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=84486a41f81c,34373,1733305383783 2024-12-04T09:43:04,735 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-04T09:43:04,736 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-04T09:43:04,794 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/hbase.id] with ID: 4d39498f-23cd-46b1-ae50-eca911105421 2024-12-04T09:43:04,794 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/.tmp/hbase.id 2024-12-04T09:43:04,800 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:04,801 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:04,806 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:47824 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:42541:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47824 dst: /127.0.0.1:42541 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:04,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-04T09:43:04,814 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:04,814 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/.tmp/hbase.id]:[hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/hbase.id] 2024-12-04T09:43:04,854 INFO [master/84486a41f81c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:04,858 INFO [master/84486a41f81c:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-04T09:43:04,875 INFO [master/84486a41f81c:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 15ms. 2024-12-04T09:43:04,887 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:04,887 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:04,887 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:04,887 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:04,900 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:04,900 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:04,902 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:53192 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53192 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:04,906 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-04T09:43:04,907 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:04,920 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-04T09:43:04,921 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-04T09:43:04,926 INFO [master/84486a41f81c:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-04T09:43:04,951 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:04,951 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:04,954 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:47858 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:42541:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47858 dst: /127.0.0.1:42541 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:04,959 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-04T09:43:04,960 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:04,977 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store 2024-12-04T09:43:04,993 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:04,994 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:05,000 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:60258 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:34441:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:60258 dst: /127.0.0.1:34441 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:05,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-04T09:43:05,005 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:05,009 INFO [master/84486a41f81c:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-04T09:43:05,013 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:05,015 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-04T09:43:05,015 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:05,015 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:05,017 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-04T09:43:05,017 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:05,017 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:05,018 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733305385014Disabling compacts and flushes for region at 1733305385014Disabling writes for close at 1733305385017 (+3 ms)Writing region close event to WAL at 1733305385017Closed at 1733305385017 2024-12-04T09:43:05,020 WARN [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/.initializing 2024-12-04T09:43:05,020 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/WALs/84486a41f81c,34373,1733305383783 2024-12-04T09:43:05,028 INFO [master/84486a41f81c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-04T09:43:05,042 INFO [master/84486a41f81c:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C34373%2C1733305383783, suffix=, logDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/WALs/84486a41f81c,34373,1733305383783, archiveDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/oldWALs, maxLogs=10 2024-12-04T09:43:05,076 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/WALs/84486a41f81c,34373,1733305383783/84486a41f81c%2C34373%2C1733305383783.1733305385047, exclude list is [], retry=0 2024-12-04T09:43:05,092 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:05,094 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42541,DS-1df1c0e0-adcb-42a3-8d8b-7e6526c2c617,DISK] 2024-12-04T09:43:05,094 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2e03073c-8b4e-4ef7-85e4-776e29ea76bb,DISK] 2024-12-04T09:43:05,094 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34441,DS-57987938-e5b7-4626-9be2-fc660a712f31,DISK] 2024-12-04T09:43:05,096 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-04T09:43:05,132 INFO [master/84486a41f81c:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/WALs/84486a41f81c,34373,1733305383783/84486a41f81c%2C34373%2C1733305383783.1733305385047 2024-12-04T09:43:05,133 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:37275:37275),(127.0.0.1/127.0.0.1:42869:42869),(127.0.0.1/127.0.0.1:35933:35933)] 2024-12-04T09:43:05,134 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-04T09:43:05,134 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:05,138 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,138 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,170 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,191 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-04T09:43:05,194 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,196 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:05,196 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,199 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-04T09:43:05,200 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,201 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-04T09:43:05,201 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,204 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-04T09:43:05,204 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,206 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-04T09:43:05,206 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,209 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-04T09:43:05,209 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,210 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-04T09:43:05,210 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,214 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,215 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,219 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,220 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,222 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-04T09:43:05,226 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:05,231 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-04T09:43:05,233 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66794613, jitterRate=-0.00468270480632782}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-04T09:43:05,241 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733305385149Initializing all the Stores at 1733305385151 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305385152 (+1 ms)Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305385152Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305385152Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305385152Cleaning up temporary data from old regions at 1733305385220 (+68 ms)Region opened successfully at 1733305385241 (+21 ms) 2024-12-04T09:43:05,243 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-04T09:43:05,271 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@ec1329b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=84486a41f81c/172.17.0.2:0 2024-12-04T09:43:05,297 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-04T09:43:05,307 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-04T09:43:05,307 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-04T09:43:05,310 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-04T09:43:05,311 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-04T09:43:05,317 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 5 msec 2024-12-04T09:43:05,318 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-04T09:43:05,344 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-04T09:43:05,352 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-04T09:43:05,395 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-04T09:43:05,398 INFO [master/84486a41f81c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-04T09:43:05,400 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-04T09:43:05,406 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-04T09:43:05,409 INFO [master/84486a41f81c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-04T09:43:05,413 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-04T09:43:05,423 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-04T09:43:05,424 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-04T09:43:05,431 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-04T09:43:05,451 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-04T09:43:05,461 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-04T09:43:05,473 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:05,473 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:05,474 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,479 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=84486a41f81c,34373,1733305383783, sessionid=0x101a102daf10000, setting cluster-up flag (Was=false) 2024-12-04T09:43:05,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,503 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,531 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-04T09:43:05,536 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=84486a41f81c,34373,1733305383783 2024-12-04T09:43:05,556 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,556 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,556 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,557 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:05,581 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-04T09:43:05,585 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=84486a41f81c,34373,1733305383783 2024-12-04T09:43:05,594 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-04T09:43:05,657 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(746): ClusterId : 4d39498f-23cd-46b1-ae50-eca911105421 2024-12-04T09:43:05,657 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(746): ClusterId : 4d39498f-23cd-46b1-ae50-eca911105421 2024-12-04T09:43:05,657 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(746): ClusterId : 4d39498f-23cd-46b1-ae50-eca911105421 2024-12-04T09:43:05,659 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-04T09:43:05,660 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-04T09:43:05,660 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-04T09:43:05,660 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-04T09:43:05,667 INFO [master/84486a41f81c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-04T09:43:05,673 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-04T09:43:05,673 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-04T09:43:05,673 INFO [master/84486a41f81c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-04T09:43:05,673 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-04T09:43:05,673 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-04T09:43:05,673 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-04T09:43:05,674 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-04T09:43:05,678 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 84486a41f81c,34373,1733305383783 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-04T09:43:05,691 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-04T09:43:05,691 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-04T09:43:05,691 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-04T09:43:05,691 DEBUG [RS:1;84486a41f81c:34771 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@68f0ddf4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=84486a41f81c/172.17.0.2:0 2024-12-04T09:43:05,691 DEBUG [RS:0;84486a41f81c:40823 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1c9f8e2c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=84486a41f81c/172.17.0.2:0 2024-12-04T09:43:05,691 DEBUG [RS:2;84486a41f81c:39799 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6d370a27, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=84486a41f81c/172.17.0.2:0 2024-12-04T09:43:05,693 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/84486a41f81c:0, corePoolSize=5, maxPoolSize=5 2024-12-04T09:43:05,693 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/84486a41f81c:0, corePoolSize=5, maxPoolSize=5 2024-12-04T09:43:05,693 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/84486a41f81c:0, corePoolSize=5, maxPoolSize=5 2024-12-04T09:43:05,693 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/84486a41f81c:0, corePoolSize=5, maxPoolSize=5 2024-12-04T09:43:05,693 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/84486a41f81c:0, corePoolSize=10, maxPoolSize=10 2024-12-04T09:43:05,694 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,694 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/84486a41f81c:0, corePoolSize=2, maxPoolSize=2 2024-12-04T09:43:05,694 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,705 DEBUG [RS:2;84486a41f81c:39799 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;84486a41f81c:39799 2024-12-04T09:43:05,708 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733305415708 2024-12-04T09:43:05,708 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-04T09:43:05,709 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-04T09:43:05,709 INFO [RS:2;84486a41f81c:39799 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-04T09:43:05,710 INFO [RS:2;84486a41f81c:39799 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-04T09:43:05,710 DEBUG [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-04T09:43:05,711 DEBUG [RS:1;84486a41f81c:34771 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;84486a41f81c:34771 2024-12-04T09:43:05,711 INFO [RS:1;84486a41f81c:34771 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-04T09:43:05,712 INFO [RS:1;84486a41f81c:34771 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-04T09:43:05,712 DEBUG [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-04T09:43:05,712 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-04T09:43:05,713 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(2659): reportForDuty to master=84486a41f81c,34373,1733305383783 with port=39799, startcode=1733305384625 2024-12-04T09:43:05,713 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(2659): reportForDuty to master=84486a41f81c,34373,1733305383783 with port=34771, startcode=1733305384579 2024-12-04T09:43:05,714 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-04T09:43:05,714 DEBUG [RS:0;84486a41f81c:40823 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;84486a41f81c:40823 2024-12-04T09:43:05,714 INFO [RS:0;84486a41f81c:40823 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-04T09:43:05,715 INFO [RS:0;84486a41f81c:40823 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-04T09:43:05,715 DEBUG [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-04T09:43:05,716 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(2659): reportForDuty to master=84486a41f81c,34373,1733305383783 with port=40823, startcode=1733305384489 2024-12-04T09:43:05,717 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,718 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-04T09:43:05,717 INFO [PEWorker-2 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-04T09:43:05,718 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-04T09:43:05,718 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-04T09:43:05,718 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-04T09:43:05,722 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,727 DEBUG [RS:1;84486a41f81c:34771 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-04T09:43:05,727 DEBUG [RS:0;84486a41f81c:40823 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-04T09:43:05,727 DEBUG [RS:2;84486a41f81c:39799 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-04T09:43:05,728 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-04T09:43:05,730 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-04T09:43:05,730 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-04T09:43:05,733 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-04T09:43:05,733 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-04T09:43:05,734 WARN [PEWorker-2 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:05,734 WARN [PEWorker-2 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:05,742 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.large.0-1733305385734,5,FailOnTimeoutGroup] 2024-12-04T09:43:05,747 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.small.0-1733305385743,5,FailOnTimeoutGroup] 2024-12-04T09:43:05,747 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,747 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-04T09:43:05,749 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,749 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,759 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:53220 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53220 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:05,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-04T09:43:05,773 WARN [PEWorker-2 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:05,773 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52007, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-04T09:43:05,773 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37769, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-04T09:43:05,773 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60951, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-04T09:43:05,776 INFO [PEWorker-2 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-04T09:43:05,776 INFO [PEWorker-2 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241 2024-12-04T09:43:05,784 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34373 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 84486a41f81c,39799,1733305384625 2024-12-04T09:43:05,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-04T09:43:05,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-04T09:43:05,788 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34373 {}] master.ServerManager(517): Registering regionserver=84486a41f81c,39799,1733305384625 2024-12-04T09:43:05,792 WARN [PEWorker-2 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:05,793 WARN [PEWorker-2 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:05,800 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:47896 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:42541:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47896 dst: /127.0.0.1:42541 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:05,801 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34373 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 84486a41f81c,34771,1733305384579 2024-12-04T09:43:05,801 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34373 {}] master.ServerManager(517): Registering regionserver=84486a41f81c,34771,1733305384579 2024-12-04T09:43:05,805 DEBUG [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241 2024-12-04T09:43:05,806 DEBUG [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:37647 2024-12-04T09:43:05,806 DEBUG [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-04T09:43:05,806 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34373 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 84486a41f81c,40823,1733305384489 2024-12-04T09:43:05,807 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34373 {}] master.ServerManager(517): Registering regionserver=84486a41f81c,40823,1733305384489 2024-12-04T09:43:05,807 DEBUG [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241 2024-12-04T09:43:05,807 DEBUG [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:37647 2024-12-04T09:43:05,807 DEBUG [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-04T09:43:05,810 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-04T09:43:05,812 DEBUG [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241 2024-12-04T09:43:05,812 WARN [PEWorker-2 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:05,812 DEBUG [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:37647 2024-12-04T09:43:05,812 DEBUG [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-04T09:43:05,813 DEBUG [PEWorker-2 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:05,817 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-04T09:43:05,819 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-04T09:43:05,819 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,820 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:05,821 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-04T09:43:05,823 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-04T09:43:05,823 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,824 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:05,824 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-04T09:43:05,827 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-04T09:43:05,827 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,828 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:05,828 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-04T09:43:05,831 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-04T09:43:05,832 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:05,833 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:05,833 DEBUG [PEWorker-2 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-04T09:43:05,835 DEBUG [PEWorker-2 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740 2024-12-04T09:43:05,835 DEBUG [PEWorker-2 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740 2024-12-04T09:43:05,838 DEBUG [PEWorker-2 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-04T09:43:05,838 DEBUG [PEWorker-2 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-04T09:43:05,839 DEBUG [PEWorker-2 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-04T09:43:05,841 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:05,842 DEBUG [PEWorker-2 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-04T09:43:05,852 DEBUG [PEWorker-2 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-04T09:43:05,853 INFO [PEWorker-2 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65464946, jitterRate=-0.02449628710746765}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-04T09:43:05,855 DEBUG [PEWorker-2 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733305385814Initializing all the Stores at 1733305385816 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305385816Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305385816Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305385816Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305385816Cleaning up temporary data from old regions at 1733305385838 (+22 ms)Region opened successfully at 1733305385855 (+17 ms) 2024-12-04T09:43:05,855 DEBUG [PEWorker-2 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-04T09:43:05,855 INFO [PEWorker-2 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-04T09:43:05,855 DEBUG [PEWorker-2 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-04T09:43:05,855 DEBUG [PEWorker-2 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-04T09:43:05,855 DEBUG [PEWorker-2 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-04T09:43:05,856 DEBUG [RS:2;84486a41f81c:39799 {}] zookeeper.ZKUtil(111): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/84486a41f81c,39799,1733305384625 2024-12-04T09:43:05,856 DEBUG [RS:1;84486a41f81c:34771 {}] zookeeper.ZKUtil(111): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/84486a41f81c,34771,1733305384579 2024-12-04T09:43:05,857 WARN [RS:2;84486a41f81c:39799 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-04T09:43:05,857 WARN [RS:1;84486a41f81c:34771 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-04T09:43:05,857 INFO [RS:2;84486a41f81c:39799 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-04T09:43:05,857 INFO [RS:1;84486a41f81c:34771 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-04T09:43:05,857 DEBUG [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,39799,1733305384625 2024-12-04T09:43:05,857 DEBUG [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,34771,1733305384579 2024-12-04T09:43:05,857 DEBUG [RS:0;84486a41f81c:40823 {}] zookeeper.ZKUtil(111): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/84486a41f81c,40823,1733305384489 2024-12-04T09:43:05,857 WARN [RS:0;84486a41f81c:40823 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-04T09:43:05,857 INFO [RS:0;84486a41f81c:40823 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-04T09:43:05,858 DEBUG [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,40823,1733305384489 2024-12-04T09:43:05,858 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [84486a41f81c,40823,1733305384489] 2024-12-04T09:43:05,859 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [84486a41f81c,34771,1733305384579] 2024-12-04T09:43:05,859 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [84486a41f81c,39799,1733305384625] 2024-12-04T09:43:05,860 INFO [PEWorker-2 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-04T09:43:05,860 DEBUG [PEWorker-2 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733305385855Disabling compacts and flushes for region at 1733305385855Disabling writes for close at 1733305385855Writing region close event to WAL at 1733305385859 (+4 ms)Closed at 1733305385860 (+1 ms) 2024-12-04T09:43:05,864 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-04T09:43:05,864 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-04T09:43:05,873 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-04T09:43:05,884 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-04T09:43:05,886 INFO [RS:2;84486a41f81c:39799 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-04T09:43:05,886 INFO [RS:0;84486a41f81c:40823 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-04T09:43:05,886 INFO [RS:1;84486a41f81c:34771 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-04T09:43:05,888 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-04T09:43:05,900 INFO [RS:2;84486a41f81c:39799 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-04T09:43:05,900 INFO [RS:0;84486a41f81c:40823 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-04T09:43:05,900 INFO [RS:1;84486a41f81c:34771 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-04T09:43:05,906 INFO [RS:2;84486a41f81c:39799 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-04T09:43:05,906 INFO [RS:0;84486a41f81c:40823 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-04T09:43:05,906 INFO [RS:1;84486a41f81c:34771 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-04T09:43:05,906 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,906 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,906 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,907 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-04T09:43:05,907 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-04T09:43:05,907 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-04T09:43:05,915 INFO [RS:0;84486a41f81c:40823 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-04T09:43:05,915 INFO [RS:2;84486a41f81c:39799 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-04T09:43:05,915 INFO [RS:1;84486a41f81c:34771 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-04T09:43:05,917 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,917 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,917 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,917 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,917 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,917 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,917 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,917 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/84486a41f81c:0, corePoolSize=2, maxPoolSize=2 2024-12-04T09:43:05,918 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/84486a41f81c:0, corePoolSize=2, maxPoolSize=2 2024-12-04T09:43:05,918 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/84486a41f81c:0, corePoolSize=2, maxPoolSize=2 2024-12-04T09:43:05,918 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,918 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:05,919 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:05,919 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:05,919 DEBUG [RS:1;84486a41f81c:34771 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:05,919 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:05,919 DEBUG [RS:2;84486a41f81c:39799 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:05,919 DEBUG [RS:0;84486a41f81c:40823 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:05,922 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,922 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,922 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,922 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,922 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,922 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,39799,1733305384625-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,40823,1733305384489-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,923 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,34771,1733305384579-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-04T09:43:05,944 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-04T09:43:05,944 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-04T09:43:05,946 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,40823,1733305384489-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,946 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,39799,1733305384625-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,946 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,946 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,947 INFO [RS:0;84486a41f81c:40823 {}] regionserver.Replication(171): 84486a41f81c,40823,1733305384489 started 2024-12-04T09:43:05,947 INFO [RS:2;84486a41f81c:39799 {}] regionserver.Replication(171): 84486a41f81c,39799,1733305384625 started 2024-12-04T09:43:05,950 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-04T09:43:05,950 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,34771,1733305384579-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,950 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,950 INFO [RS:1;84486a41f81c:34771 {}] regionserver.Replication(171): 84486a41f81c,34771,1733305384579 started 2024-12-04T09:43:05,966 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,967 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1482): Serving as 84486a41f81c,40823,1733305384489, RpcServer on 84486a41f81c/172.17.0.2:40823, sessionid=0x101a102daf10001 2024-12-04T09:43:05,968 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-04T09:43:05,968 DEBUG [RS:0;84486a41f81c:40823 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 84486a41f81c,40823,1733305384489 2024-12-04T09:43:05,968 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,40823,1733305384489' 2024-12-04T09:43:05,969 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-04T09:43:05,970 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-04T09:43:05,970 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,970 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1482): Serving as 84486a41f81c,39799,1733305384625, RpcServer on 84486a41f81c/172.17.0.2:39799, sessionid=0x101a102daf10003 2024-12-04T09:43:05,971 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-04T09:43:05,971 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-04T09:43:05,971 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-04T09:43:05,971 DEBUG [RS:2;84486a41f81c:39799 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 84486a41f81c,39799,1733305384625 2024-12-04T09:43:05,971 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,39799,1733305384625' 2024-12-04T09:43:05,971 DEBUG [RS:0;84486a41f81c:40823 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 84486a41f81c,40823,1733305384489 2024-12-04T09:43:05,971 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-04T09:43:05,971 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,40823,1733305384489' 2024-12-04T09:43:05,971 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-04T09:43:05,972 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-04T09:43:05,972 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:05,972 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-04T09:43:05,972 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(1482): Serving as 84486a41f81c,34771,1733305384579, RpcServer on 84486a41f81c/172.17.0.2:34771, sessionid=0x101a102daf10002 2024-12-04T09:43:05,972 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-04T09:43:05,973 DEBUG [RS:1;84486a41f81c:34771 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 84486a41f81c,34771,1733305384579 2024-12-04T09:43:05,973 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,34771,1733305384579' 2024-12-04T09:43:05,973 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-04T09:43:05,973 DEBUG [RS:0;84486a41f81c:40823 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-04T09:43:05,973 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-04T09:43:05,973 INFO [RS:0;84486a41f81c:40823 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-04T09:43:05,973 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-04T09:43:05,973 INFO [RS:0;84486a41f81c:40823 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-04T09:43:05,973 DEBUG [RS:2;84486a41f81c:39799 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 84486a41f81c,39799,1733305384625 2024-12-04T09:43:05,973 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,39799,1733305384625' 2024-12-04T09:43:05,973 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-04T09:43:05,973 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-04T09:43:05,974 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-04T09:43:05,974 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-04T09:43:05,974 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-04T09:43:05,974 DEBUG [RS:1;84486a41f81c:34771 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 84486a41f81c,34771,1733305384579 2024-12-04T09:43:05,974 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,34771,1733305384579' 2024-12-04T09:43:05,974 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-04T09:43:05,974 DEBUG [RS:2;84486a41f81c:39799 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-04T09:43:05,974 INFO [RS:2;84486a41f81c:39799 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-04T09:43:05,974 INFO [RS:2;84486a41f81c:39799 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-04T09:43:05,975 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-04T09:43:05,975 DEBUG [RS:1;84486a41f81c:34771 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-04T09:43:05,975 INFO [RS:1;84486a41f81c:34771 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-04T09:43:05,975 INFO [RS:1;84486a41f81c:34771 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-04T09:43:06,039 WARN [84486a41f81c:34373 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-04T09:43:06,079 INFO [RS:0;84486a41f81c:40823 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-04T09:43:06,079 INFO [RS:1;84486a41f81c:34771 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-04T09:43:06,079 INFO [RS:2;84486a41f81c:39799 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-04T09:43:06,083 INFO [RS:0;84486a41f81c:40823 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C40823%2C1733305384489, suffix=, logDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,40823,1733305384489, archiveDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/oldWALs, maxLogs=32 2024-12-04T09:43:06,083 INFO [RS:1;84486a41f81c:34771 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C34771%2C1733305384579, suffix=, logDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,34771,1733305384579, archiveDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/oldWALs, maxLogs=32 2024-12-04T09:43:06,083 INFO [RS:2;84486a41f81c:39799 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C39799%2C1733305384625, suffix=, logDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,39799,1733305384625, archiveDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/oldWALs, maxLogs=32 2024-12-04T09:43:06,097 DEBUG [RS:1;84486a41f81c:34771 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,34771,1733305384579/84486a41f81c%2C34771%2C1733305384579.1733305386086, exclude list is [], retry=0 2024-12-04T09:43:06,097 DEBUG [RS:0;84486a41f81c:40823 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,40823,1733305384489/84486a41f81c%2C40823%2C1733305384489.1733305386086, exclude list is [], retry=0 2024-12-04T09:43:06,099 DEBUG [RS:2;84486a41f81c:39799 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,39799,1733305384625/84486a41f81c%2C39799%2C1733305384625.1733305386086, exclude list is [], retry=0 2024-12-04T09:43:06,102 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42541,DS-1df1c0e0-adcb-42a3-8d8b-7e6526c2c617,DISK] 2024-12-04T09:43:06,102 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2e03073c-8b4e-4ef7-85e4-776e29ea76bb,DISK] 2024-12-04T09:43:06,103 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34441,DS-57987938-e5b7-4626-9be2-fc660a712f31,DISK] 2024-12-04T09:43:06,103 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42541,DS-1df1c0e0-adcb-42a3-8d8b-7e6526c2c617,DISK] 2024-12-04T09:43:06,103 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34441,DS-57987938-e5b7-4626-9be2-fc660a712f31,DISK] 2024-12-04T09:43:06,104 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2e03073c-8b4e-4ef7-85e4-776e29ea76bb,DISK] 2024-12-04T09:43:06,123 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2e03073c-8b4e-4ef7-85e4-776e29ea76bb,DISK] 2024-12-04T09:43:06,123 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34441,DS-57987938-e5b7-4626-9be2-fc660a712f31,DISK] 2024-12-04T09:43:06,124 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42541,DS-1df1c0e0-adcb-42a3-8d8b-7e6526c2c617,DISK] 2024-12-04T09:43:06,133 INFO [RS:1;84486a41f81c:34771 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,34771,1733305384579/84486a41f81c%2C34771%2C1733305384579.1733305386086 2024-12-04T09:43:06,134 INFO [RS:0;84486a41f81c:40823 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,40823,1733305384489/84486a41f81c%2C40823%2C1733305384489.1733305386086 2024-12-04T09:43:06,135 DEBUG [RS:1;84486a41f81c:34771 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:35933:35933),(127.0.0.1/127.0.0.1:37275:37275),(127.0.0.1/127.0.0.1:42869:42869)] 2024-12-04T09:43:06,135 INFO [RS:2;84486a41f81c:39799 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,39799,1733305384625/84486a41f81c%2C39799%2C1733305384625.1733305386086 2024-12-04T09:43:06,135 DEBUG [RS:0;84486a41f81c:40823 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:42869:42869),(127.0.0.1/127.0.0.1:35933:35933),(127.0.0.1/127.0.0.1:37275:37275)] 2024-12-04T09:43:06,137 DEBUG [RS:2;84486a41f81c:39799 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:42869:42869),(127.0.0.1/127.0.0.1:37275:37275),(127.0.0.1/127.0.0.1:35933:35933)] 2024-12-04T09:43:06,291 DEBUG [84486a41f81c:34373 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-04T09:43:06,301 DEBUG [84486a41f81c:34373 {}] balancer.BalancerClusterState(204): Hosts are {84486a41f81c=0} racks are {/default-rack=0} 2024-12-04T09:43:06,307 DEBUG [84486a41f81c:34373 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-04T09:43:06,307 DEBUG [84486a41f81c:34373 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-04T09:43:06,307 DEBUG [84486a41f81c:34373 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-04T09:43:06,307 DEBUG [84486a41f81c:34373 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-04T09:43:06,307 DEBUG [84486a41f81c:34373 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-04T09:43:06,307 DEBUG [84486a41f81c:34373 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-04T09:43:06,307 INFO [84486a41f81c:34373 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-04T09:43:06,307 INFO [84486a41f81c:34373 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-04T09:43:06,307 INFO [84486a41f81c:34373 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-04T09:43:06,307 DEBUG [84486a41f81c:34373 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-04T09:43:06,313 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=84486a41f81c,40823,1733305384489 2024-12-04T09:43:06,318 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 84486a41f81c,40823,1733305384489, state=OPENING 2024-12-04T09:43:06,356 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-04T09:43:06,365 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:06,365 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:06,365 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:06,365 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:06,365 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:06,365 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:06,366 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:06,366 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:06,367 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-04T09:43:06,369 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=84486a41f81c,40823,1733305384489}] 2024-12-04T09:43:06,549 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-04T09:43:06,551 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43251, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-04T09:43:06,562 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-04T09:43:06,563 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-04T09:43:06,563 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-04T09:43:06,567 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C40823%2C1733305384489.meta, suffix=.meta, logDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,40823,1733305384489, archiveDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/oldWALs, maxLogs=32 2024-12-04T09:43:06,581 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,40823,1733305384489/84486a41f81c%2C40823%2C1733305384489.meta.1733305386568.meta, exclude list is [], retry=0 2024-12-04T09:43:06,584 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2e03073c-8b4e-4ef7-85e4-776e29ea76bb,DISK] 2024-12-04T09:43:06,584 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42541,DS-1df1c0e0-adcb-42a3-8d8b-7e6526c2c617,DISK] 2024-12-04T09:43:06,584 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34441,DS-57987938-e5b7-4626-9be2-fc660a712f31,DISK] 2024-12-04T09:43:06,588 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,40823,1733305384489/84486a41f81c%2C40823%2C1733305384489.meta.1733305386568.meta 2024-12-04T09:43:06,589 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:42869:42869),(127.0.0.1/127.0.0.1:35933:35933),(127.0.0.1/127.0.0.1:37275:37275)] 2024-12-04T09:43:06,589 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-04T09:43:06,591 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-04T09:43:06,593 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-04T09:43:06,597 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-04T09:43:06,600 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-04T09:43:06,601 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:06,601 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-04T09:43:06,601 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-04T09:43:06,604 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-04T09:43:06,605 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-04T09:43:06,605 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:06,606 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:06,606 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-04T09:43:06,608 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-04T09:43:06,608 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:06,609 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:06,609 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-04T09:43:06,610 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-04T09:43:06,610 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:06,611 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:06,611 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-04T09:43:06,612 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-04T09:43:06,612 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:06,613 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:06,614 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-04T09:43:06,615 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740 2024-12-04T09:43:06,619 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740 2024-12-04T09:43:06,621 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-04T09:43:06,621 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-04T09:43:06,622 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-04T09:43:06,625 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-04T09:43:06,626 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74991276, jitterRate=0.11745709180831909}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-04T09:43:06,626 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-04T09:43:06,628 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733305386601Writing region info on filesystem at 1733305386601Initializing all the Stores at 1733305386603 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305386603Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305386604 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305386604Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305386604Cleaning up temporary data from old regions at 1733305386621 (+17 ms)Running coprocessor post-open hooks at 1733305386626 (+5 ms)Region opened successfully at 1733305386628 (+2 ms) 2024-12-04T09:43:06,635 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733305386542 2024-12-04T09:43:06,644 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-04T09:43:06,645 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-04T09:43:06,646 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=84486a41f81c,40823,1733305384489 2024-12-04T09:43:06,648 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 84486a41f81c,40823,1733305384489, state=OPEN 2024-12-04T09:43:06,656 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-04T09:43:06,656 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-04T09:43:06,656 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-04T09:43:06,656 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-04T09:43:06,656 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:06,656 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:06,656 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:06,656 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:06,657 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=84486a41f81c,40823,1733305384489 2024-12-04T09:43:06,662 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-04T09:43:06,663 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=84486a41f81c,40823,1733305384489 in 288 msec 2024-12-04T09:43:06,670 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-04T09:43:06,670 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 792 msec 2024-12-04T09:43:06,671 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-04T09:43:06,672 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-04T09:43:06,688 DEBUG [PEWorker-1 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-04T09:43:06,689 DEBUG [PEWorker-1 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=84486a41f81c,40823,1733305384489, seqNum=-1] 2024-12-04T09:43:06,705 DEBUG [PEWorker-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-04T09:43:06,707 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54459, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-04T09:43:06,747 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1250 sec 2024-12-04T09:43:06,748 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733305386748, completionTime=-1 2024-12-04T09:43:06,751 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-04T09:43:06,751 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-04T09:43:06,783 INFO [master/84486a41f81c:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-04T09:43:06,783 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733305446783 2024-12-04T09:43:06,783 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733305506783 2024-12-04T09:43:06,783 INFO [master/84486a41f81c:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 32 msec 2024-12-04T09:43:06,792 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-04T09:43:06,798 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,34373,1733305383783-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:06,798 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,34373,1733305383783-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:06,798 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,34373,1733305383783-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:06,799 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-84486a41f81c:34373, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:06,799 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:06,800 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:06,805 DEBUG [master/84486a41f81c:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-04T09:43:06,826 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.092sec 2024-12-04T09:43:06,827 INFO [master/84486a41f81c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-04T09:43:06,828 INFO [master/84486a41f81c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-04T09:43:06,829 INFO [master/84486a41f81c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-04T09:43:06,829 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-04T09:43:06,829 INFO [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-04T09:43:06,830 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,34373,1733305383783-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-04T09:43:06,830 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,34373,1733305383783-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-04T09:43:06,834 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-04T09:43:06,835 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-04T09:43:06,835 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,34373,1733305383783-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:06,871 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7c2b8794, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-04T09:43:06,876 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-04T09:43:06,876 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-04T09:43:06,881 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 84486a41f81c,34373,-1 for getting cluster id 2024-12-04T09:43:06,884 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-04T09:43:06,891 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '4d39498f-23cd-46b1-ae50-eca911105421' 2024-12-04T09:43:06,893 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-04T09:43:06,893 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "4d39498f-23cd-46b1-ae50-eca911105421" 2024-12-04T09:43:06,894 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2c94782c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-04T09:43:06,894 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [84486a41f81c,34373,-1] 2024-12-04T09:43:06,896 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-04T09:43:06,898 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:06,899 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41878, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-04T09:43:06,902 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5260075b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-04T09:43:06,903 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-04T09:43:06,909 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=84486a41f81c,40823,1733305384489, seqNum=-1] 2024-12-04T09:43:06,910 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-04T09:43:06,913 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52128, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-04T09:43:06,938 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=84486a41f81c,34373,1733305383783 2024-12-04T09:43:06,943 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-04T09:43:06,949 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 84486a41f81c,34373,1733305383783 2024-12-04T09:43:06,951 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@2c5300f3 2024-12-04T09:43:06,952 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-04T09:43:06,954 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41888, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-04T09:43:06,960 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-04T09:43:06,967 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-04T09:43:06,970 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-04T09:43:06,972 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-04T09:43:06,973 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:06,975 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-04T09:43:06,978 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-04T09:43:06,985 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:06,985 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:06,988 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:47960 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:42541:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47960 dst: /127.0.0.1:42541 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:06,992 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-04T09:43:06,993 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:06,996 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 0a60f2d967287a2b25098c3876580a2a, NAME => 'TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241 2024-12-04T09:43:07,002 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:07,002 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:07,008 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:47974 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:42541:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47974 dst: /127.0.0.1:42541 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:07,012 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-04T09:43:07,013 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:07,013 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:07,013 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 0a60f2d967287a2b25098c3876580a2a, disabling compactions & flushes 2024-12-04T09:43:07,014 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,014 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,014 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. after waiting 0 ms 2024-12-04T09:43:07,014 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,014 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,014 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 0a60f2d967287a2b25098c3876580a2a: Waiting for close lock at 1733305387013Disabling compacts and flushes for region at 1733305387013Disabling writes for close at 1733305387014 (+1 ms)Writing region close event to WAL at 1733305387014Closed at 1733305387014 2024-12-04T09:43:07,016 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-04T09:43:07,020 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733305387016"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733305387016"}]},"ts":"1733305387016"} 2024-12-04T09:43:07,024 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-04T09:43:07,026 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-04T09:43:07,028 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733305387026"}]},"ts":"1733305387026"} 2024-12-04T09:43:07,032 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-04T09:43:07,032 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {84486a41f81c=0} racks are {/default-rack=0} 2024-12-04T09:43:07,034 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-04T09:43:07,034 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-04T09:43:07,034 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-04T09:43:07,034 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-04T09:43:07,034 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-04T09:43:07,034 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-04T09:43:07,034 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-04T09:43:07,034 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-04T09:43:07,034 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-04T09:43:07,034 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-04T09:43:07,035 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0a60f2d967287a2b25098c3876580a2a, ASSIGN}] 2024-12-04T09:43:07,038 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0a60f2d967287a2b25098c3876580a2a, ASSIGN 2024-12-04T09:43:07,039 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0a60f2d967287a2b25098c3876580a2a, ASSIGN; state=OFFLINE, location=84486a41f81c,39799,1733305384625; forceNewPlan=false, retain=false 2024-12-04T09:43:07,088 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-04T09:43:07,194 INFO [84486a41f81c:34373 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-04T09:43:07,195 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=0a60f2d967287a2b25098c3876580a2a, regionState=OPENING, regionLocation=84486a41f81c,39799,1733305384625 2024-12-04T09:43:07,201 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0a60f2d967287a2b25098c3876580a2a, ASSIGN because future has completed 2024-12-04T09:43:07,202 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 0a60f2d967287a2b25098c3876580a2a, server=84486a41f81c,39799,1733305384625}] 2024-12-04T09:43:07,297 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-04T09:43:07,358 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-04T09:43:07,363 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41407, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-04T09:43:07,372 INFO [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,373 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 0a60f2d967287a2b25098c3876580a2a, NAME => 'TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a.', STARTKEY => '', ENDKEY => ''} 2024-12-04T09:43:07,373 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,373 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:07,373 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,373 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,376 INFO [StoreOpener-0a60f2d967287a2b25098c3876580a2a-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,378 INFO [StoreOpener-0a60f2d967287a2b25098c3876580a2a-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 0a60f2d967287a2b25098c3876580a2a columnFamilyName cf 2024-12-04T09:43:07,378 DEBUG [StoreOpener-0a60f2d967287a2b25098c3876580a2a-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:07,379 INFO [StoreOpener-0a60f2d967287a2b25098c3876580a2a-1 {}] regionserver.HStore(327): Store=0a60f2d967287a2b25098c3876580a2a/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-04T09:43:07,379 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,380 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,381 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,382 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,382 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,384 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,389 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-04T09:43:07,390 INFO [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 0a60f2d967287a2b25098c3876580a2a; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61786975, jitterRate=-0.07930232584476471}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-04T09:43:07,390 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:07,391 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 0a60f2d967287a2b25098c3876580a2a: Running coprocessor pre-open hook at 1733305387374Writing region info on filesystem at 1733305387374Initializing all the Stores at 1733305387375 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305387375Cleaning up temporary data from old regions at 1733305387382 (+7 ms)Running coprocessor post-open hooks at 1733305387390 (+8 ms)Region opened successfully at 1733305387391 (+1 ms) 2024-12-04T09:43:07,393 INFO [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a., pid=6, masterSystemTime=1733305387357 2024-12-04T09:43:07,396 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,396 INFO [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,397 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=0a60f2d967287a2b25098c3876580a2a, regionState=OPEN, openSeqNum=2, regionLocation=84486a41f81c,39799,1733305384625 2024-12-04T09:43:07,401 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 0a60f2d967287a2b25098c3876580a2a, server=84486a41f81c,39799,1733305384625 because future has completed 2024-12-04T09:43:07,406 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-04T09:43:07,406 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 0a60f2d967287a2b25098c3876580a2a, server=84486a41f81c,39799,1733305384625 in 201 msec 2024-12-04T09:43:07,410 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-04T09:43:07,410 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0a60f2d967287a2b25098c3876580a2a, ASSIGN in 371 msec 2024-12-04T09:43:07,412 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-04T09:43:07,412 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733305387412"}]},"ts":"1733305387412"} 2024-12-04T09:43:07,415 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-04T09:43:07,416 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-04T09:43:07,419 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 454 msec 2024-12-04T09:43:07,607 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-04T09:43:07,608 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-04T09:43:07,608 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-04T09:43:07,610 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-04T09:43:07,619 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-04T09:43:07,619 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-04T09:43:07,619 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-04T09:43:07,627 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a., hostname=84486a41f81c,39799,1733305384625, seqNum=2] 2024-12-04T09:43:07,628 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-04T09:43:07,630 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43190, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-04T09:43:07,639 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-04T09:43:07,645 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-04T09:43:07,647 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-04T09:43:07,647 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-04T09:43:07,649 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-04T09:43:07,651 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-04T09:43:07,758 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-04T09:43:07,815 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39799 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-04T09:43:07,815 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,820 INFO [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 0a60f2d967287a2b25098c3876580a2a 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-04T09:43:07,870 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a/.tmp/cf/c4b6f720b4a746db828b383198195207 is 36, key is row/cf:cq/1733305387631/Put/seqid=0 2024-12-04T09:43:07,877 WARN [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:07,877 WARN [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:07,881 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_196824238_22 at /127.0.0.1:50782 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:34441:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50782 dst: /127.0.0.1:34441 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:07,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-04T09:43:07,886 WARN [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:07,886 INFO [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a/.tmp/cf/c4b6f720b4a746db828b383198195207 2024-12-04T09:43:07,927 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a/.tmp/cf/c4b6f720b4a746db828b383198195207 as hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a/cf/c4b6f720b4a746db828b383198195207 2024-12-04T09:43:07,937 INFO [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a/cf/c4b6f720b4a746db828b383198195207, entries=1, sequenceid=5, filesize=4.7 K 2024-12-04T09:43:07,943 INFO [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 0a60f2d967287a2b25098c3876580a2a in 124ms, sequenceid=5, compaction requested=false 2024-12-04T09:43:07,944 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-04T09:43:07,946 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 0a60f2d967287a2b25098c3876580a2a: 2024-12-04T09:43:07,946 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:07,947 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-04T09:43:07,949 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-04T09:43:07,955 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-04T09:43:07,955 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 300 msec 2024-12-04T09:43:07,958 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 315 msec 2024-12-04T09:43:07,966 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34373 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-04T09:43:07,967 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-04T09:43:07,980 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-04T09:43:07,980 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-04T09:43:07,980 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:07,984 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:07,984 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:07,984 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-04T09:43:07,984 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-04T09:43:07,984 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1496649153, stopped=false 2024-12-04T09:43:07,985 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=84486a41f81c,34373,1733305383783 2024-12-04T09:43:08,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:08,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:08,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:08,023 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:08,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:08,023 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:08,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:08,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:08,024 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-04T09:43:08,024 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-04T09:43:08,025 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:08,025 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:08,025 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:08,025 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:08,025 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:08,026 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '84486a41f81c,40823,1733305384489' ***** 2024-12-04T09:43:08,026 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-04T09:43:08,026 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:08,026 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '84486a41f81c,34771,1733305384579' ***** 2024-12-04T09:43:08,026 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-04T09:43:08,027 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-04T09:43:08,027 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-04T09:43:08,027 INFO [RS:1;84486a41f81c:34771 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-04T09:43:08,027 INFO [RS:0;84486a41f81c:40823 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-04T09:43:08,027 INFO [RS:1;84486a41f81c:34771 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-04T09:43:08,028 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(959): stopping server 84486a41f81c,34771,1733305384579 2024-12-04T09:43:08,028 INFO [RS:1;84486a41f81c:34771 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-04T09:43:08,028 INFO [RS:1;84486a41f81c:34771 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;84486a41f81c:34771. 2024-12-04T09:43:08,028 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-04T09:43:08,028 DEBUG [RS:1;84486a41f81c:34771 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:08,028 DEBUG [RS:1;84486a41f81c:34771 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:08,027 INFO [RS:0;84486a41f81c:40823 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-04T09:43:08,028 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(976): stopping server 84486a41f81c,34771,1733305384579; all regions closed. 2024-12-04T09:43:08,029 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(959): stopping server 84486a41f81c,40823,1733305384489 2024-12-04T09:43:08,029 INFO [RS:0;84486a41f81c:40823 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-04T09:43:08,029 INFO [RS:0;84486a41f81c:40823 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;84486a41f81c:40823. 2024-12-04T09:43:08,029 DEBUG [RS:0;84486a41f81c:40823 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:08,029 DEBUG [RS:0;84486a41f81c:40823 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:08,029 INFO [RS:0;84486a41f81c:40823 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-04T09:43:08,029 INFO [RS:0;84486a41f81c:40823 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-04T09:43:08,029 INFO [RS:0;84486a41f81c:40823 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-04T09:43:08,029 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-04T09:43:08,030 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-04T09:43:08,030 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-04T09:43:08,030 DEBUG [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-04T09:43:08,030 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '84486a41f81c,39799,1733305384625' ***** 2024-12-04T09:43:08,030 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-04T09:43:08,031 DEBUG [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-04T09:43:08,031 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-04T09:43:08,031 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-04T09:43:08,031 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-04T09:43:08,031 INFO [RS:2;84486a41f81c:39799 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-04T09:43:08,031 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-04T09:43:08,031 INFO [RS:2;84486a41f81c:39799 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-04T09:43:08,031 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-04T09:43:08,031 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(3091): Received CLOSE for 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:08,031 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-04T09:43:08,031 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-04T09:43:08,031 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(959): stopping server 84486a41f81c,39799,1733305384625 2024-12-04T09:43:08,032 INFO [RS:2;84486a41f81c:39799 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-04T09:43:08,032 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-04T09:43:08,032 INFO [regionserver/84486a41f81c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:08,032 INFO [regionserver/84486a41f81c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:08,032 INFO [RS:2;84486a41f81c:39799 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;84486a41f81c:39799. 2024-12-04T09:43:08,032 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 0a60f2d967287a2b25098c3876580a2a, disabling compactions & flushes 2024-12-04T09:43:08,032 DEBUG [RS:2;84486a41f81c:39799 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:08,032 INFO [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:08,032 DEBUG [RS:2;84486a41f81c:39799 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:08,032 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:08,032 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. after waiting 0 ms 2024-12-04T09:43:08,032 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-04T09:43:08,032 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:08,032 INFO [regionserver/84486a41f81c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:08,032 DEBUG [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1325): Online Regions={0a60f2d967287a2b25098c3876580a2a=TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a.} 2024-12-04T09:43:08,032 DEBUG [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1351): Waiting on 0a60f2d967287a2b25098c3876580a2a 2024-12-04T09:43:08,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_1073741826_1016 (size=93) 2024-12-04T09:43:08,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_1073741826_1016 (size=93) 2024-12-04T09:43:08,037 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741826_1016 (size=93) 2024-12-04T09:43:08,042 DEBUG [RS:1;84486a41f81c:34771 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/oldWALs 2024-12-04T09:43:08,043 INFO [RS:1;84486a41f81c:34771 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 84486a41f81c%2C34771%2C1733305384579:(num 1733305386086) 2024-12-04T09:43:08,043 DEBUG [RS:1;84486a41f81c:34771 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:08,043 INFO [RS:1;84486a41f81c:34771 {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:08,043 INFO [RS:1;84486a41f81c:34771 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-04T09:43:08,043 INFO [RS:1;84486a41f81c:34771 {}] hbase.ChoreService(370): Chore service for: regionserver/84486a41f81c:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-04T09:43:08,043 INFO [RS:1;84486a41f81c:34771 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-04T09:43:08,044 INFO [RS:1;84486a41f81c:34771 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-04T09:43:08,044 INFO [regionserver/84486a41f81c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-04T09:43:08,044 INFO [RS:1;84486a41f81c:34771 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-04T09:43:08,044 INFO [RS:1;84486a41f81c:34771 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-04T09:43:08,044 INFO [RS:1;84486a41f81c:34771 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:34771 2024-12-04T09:43:08,053 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/84486a41f81c,34771,1733305384579 2024-12-04T09:43:08,053 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:08,053 INFO [RS:1;84486a41f81c:34771 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-04T09:43:08,054 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [84486a41f81c,34771,1733305384579] 2024-12-04T09:43:08,055 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/default/TestHBaseWalOnEC/0a60f2d967287a2b25098c3876580a2a/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-04T09:43:08,058 INFO [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:08,058 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 0a60f2d967287a2b25098c3876580a2a: Waiting for close lock at 1733305388032Running coprocessor pre-close hooks at 1733305388032Disabling compacts and flushes for region at 1733305388032Disabling writes for close at 1733305388032Writing region close event to WAL at 1733305388034 (+2 ms)Running coprocessor post-close hooks at 1733305388056 (+22 ms)Closed at 1733305388057 (+1 ms) 2024-12-04T09:43:08,058 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a. 2024-12-04T09:43:08,064 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/info/cf20e61ff3204a86a99baf8cdb2fc227 is 153, key is TestHBaseWalOnEC,,1733305386956.0a60f2d967287a2b25098c3876580a2a./info:regioninfo/1733305387397/Put/seqid=0 2024-12-04T09:43:08,067 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,067 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,070 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/84486a41f81c,34771,1733305384579 already deleted, retry=false 2024-12-04T09:43:08,070 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 84486a41f81c,34771,1733305384579 expired; onlineServers=2 2024-12-04T09:43:08,071 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-850994977_22 at /127.0.0.1:50798 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:34441:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50798 dst: /127.0.0.1:34441 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:08,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-04T09:43:08,075 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:08,076 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/info/cf20e61ff3204a86a99baf8cdb2fc227 2024-12-04T09:43:08,101 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/ns/4a4e46260b974475954b0fe4d0f3f1e7 is 43, key is default/ns:d/1733305386731/Put/seqid=0 2024-12-04T09:43:08,103 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,103 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,106 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-850994977_22 at /127.0.0.1:34148 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34148 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:08,110 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-04T09:43:08,111 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:08,111 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/ns/4a4e46260b974475954b0fe4d0f3f1e7 2024-12-04T09:43:08,135 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/table/8dabcc3d04034c05ad4e3166f9ee99b1 is 52, key is TestHBaseWalOnEC/table:state/1733305387412/Put/seqid=0 2024-12-04T09:43:08,138 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,138 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,141 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-850994977_22 at /127.0.0.1:50818 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:34441:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50818 dst: /127.0.0.1:34441 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:08,145 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-04T09:43:08,146 WARN [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:08,146 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/table/8dabcc3d04034c05ad4e3166f9ee99b1 2024-12-04T09:43:08,156 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/info/cf20e61ff3204a86a99baf8cdb2fc227 as hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/info/cf20e61ff3204a86a99baf8cdb2fc227 2024-12-04T09:43:08,162 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:08,162 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34771-0x101a102daf10002, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:08,162 INFO [RS:1;84486a41f81c:34771 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-04T09:43:08,163 INFO [RS:1;84486a41f81c:34771 {}] regionserver.HRegionServer(1031): Exiting; stopping=84486a41f81c,34771,1733305384579; zookeeper connection closed. 2024-12-04T09:43:08,163 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@50ef224e {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@50ef224e 2024-12-04T09:43:08,168 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/info/cf20e61ff3204a86a99baf8cdb2fc227, entries=10, sequenceid=11, filesize=6.5 K 2024-12-04T09:43:08,169 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/ns/4a4e46260b974475954b0fe4d0f3f1e7 as hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/ns/4a4e46260b974475954b0fe4d0f3f1e7 2024-12-04T09:43:08,179 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/ns/4a4e46260b974475954b0fe4d0f3f1e7, entries=2, sequenceid=11, filesize=5.0 K 2024-12-04T09:43:08,181 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/.tmp/table/8dabcc3d04034c05ad4e3166f9ee99b1 as hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/table/8dabcc3d04034c05ad4e3166f9ee99b1 2024-12-04T09:43:08,191 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/table/8dabcc3d04034c05ad4e3166f9ee99b1, entries=2, sequenceid=11, filesize=5.1 K 2024-12-04T09:43:08,193 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 161ms, sequenceid=11, compaction requested=false 2024-12-04T09:43:08,193 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-04T09:43:08,201 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-04T09:43:08,202 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-04T09:43:08,203 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-04T09:43:08,203 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733305388030Running coprocessor pre-close hooks at 1733305388030Disabling compacts and flushes for region at 1733305388031 (+1 ms)Disabling writes for close at 1733305388031Obtaining lock to block concurrent updates at 1733305388032 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1733305388032Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733305388032Flushing stores of hbase:meta,,1.1588230740 at 1733305388034 (+2 ms)Flushing 1588230740/info: creating writer at 1733305388034Flushing 1588230740/info: appending metadata at 1733305388060 (+26 ms)Flushing 1588230740/info: closing flushed file at 1733305388060Flushing 1588230740/ns: creating writer at 1733305388085 (+25 ms)Flushing 1588230740/ns: appending metadata at 1733305388100 (+15 ms)Flushing 1588230740/ns: closing flushed file at 1733305388100Flushing 1588230740/table: creating writer at 1733305388118 (+18 ms)Flushing 1588230740/table: appending metadata at 1733305388133 (+15 ms)Flushing 1588230740/table: closing flushed file at 1733305388133Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@515ee98c: reopening flushed file at 1733305388154 (+21 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@216c0cbb: reopening flushed file at 1733305388168 (+14 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@46c8fada: reopening flushed file at 1733305388179 (+11 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 161ms, sequenceid=11, compaction requested=false at 1733305388193 (+14 ms)Writing region close event to WAL at 1733305388194 (+1 ms)Running coprocessor post-close hooks at 1733305388202 (+8 ms)Closed at 1733305388203 (+1 ms) 2024-12-04T09:43:08,203 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-04T09:43:08,231 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(976): stopping server 84486a41f81c,40823,1733305384489; all regions closed. 2024-12-04T09:43:08,232 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(976): stopping server 84486a41f81c,39799,1733305384625; all regions closed. 2024-12-04T09:43:08,234 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741829_1019 (size=2751) 2024-12-04T09:43:08,234 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_1073741829_1019 (size=2751) 2024-12-04T09:43:08,235 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_1073741829_1019 (size=2751) 2024-12-04T09:43:08,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_1073741828_1018 (size=1298) 2024-12-04T09:43:08,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741828_1018 (size=1298) 2024-12-04T09:43:08,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_1073741828_1018 (size=1298) 2024-12-04T09:43:08,245 DEBUG [RS:2;84486a41f81c:39799 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/oldWALs 2024-12-04T09:43:08,245 INFO [RS:2;84486a41f81c:39799 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 84486a41f81c%2C39799%2C1733305384625:(num 1733305386086) 2024-12-04T09:43:08,245 DEBUG [RS:2;84486a41f81c:39799 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:08,245 DEBUG [RS:0;84486a41f81c:40823 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/oldWALs 2024-12-04T09:43:08,245 INFO [RS:2;84486a41f81c:39799 {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:08,245 INFO [RS:0;84486a41f81c:40823 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 84486a41f81c%2C40823%2C1733305384489.meta:.meta(num 1733305386568) 2024-12-04T09:43:08,245 INFO [RS:2;84486a41f81c:39799 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-04T09:43:08,245 INFO [RS:2;84486a41f81c:39799 {}] hbase.ChoreService(370): Chore service for: regionserver/84486a41f81c:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-04T09:43:08,245 INFO [RS:2;84486a41f81c:39799 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-04T09:43:08,245 INFO [RS:2;84486a41f81c:39799 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-04T09:43:08,245 INFO [regionserver/84486a41f81c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-04T09:43:08,245 INFO [RS:2;84486a41f81c:39799 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-04T09:43:08,245 INFO [RS:2;84486a41f81c:39799 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-04T09:43:08,246 INFO [RS:2;84486a41f81c:39799 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39799 2024-12-04T09:43:08,248 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_1073741827_1017 (size=93) 2024-12-04T09:43:08,249 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741827_1017 (size=93) 2024-12-04T09:43:08,248 WARN [Close-WAL-Writer-0 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(650): complete file /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/WALs/84486a41f81c,40823,1733305384489/84486a41f81c%2C40823%2C1733305384489.1733305386086 not finished, retry = 0 2024-12-04T09:43:08,249 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_1073741827_1017 (size=93) 2024-12-04T09:43:08,256 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/84486a41f81c,39799,1733305384625 2024-12-04T09:43:08,256 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:08,256 INFO [RS:2;84486a41f81c:39799 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-04T09:43:08,257 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [84486a41f81c,39799,1733305384625] 2024-12-04T09:43:08,273 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/84486a41f81c,39799,1733305384625 already deleted, retry=false 2024-12-04T09:43:08,273 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 84486a41f81c,39799,1733305384625 expired; onlineServers=1 2024-12-04T09:43:08,356 DEBUG [RS:0;84486a41f81c:40823 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/oldWALs 2024-12-04T09:43:08,356 INFO [RS:0;84486a41f81c:40823 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 84486a41f81c%2C40823%2C1733305384489:(num 1733305386086) 2024-12-04T09:43:08,356 DEBUG [RS:0;84486a41f81c:40823 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:08,356 INFO [RS:0;84486a41f81c:40823 {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:08,356 INFO [RS:0;84486a41f81c:40823 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-04T09:43:08,357 INFO [RS:0;84486a41f81c:40823 {}] hbase.ChoreService(370): Chore service for: regionserver/84486a41f81c:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-04T09:43:08,357 INFO [RS:0;84486a41f81c:40823 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-04T09:43:08,357 INFO [regionserver/84486a41f81c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-04T09:43:08,358 INFO [RS:0;84486a41f81c:40823 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:40823 2024-12-04T09:43:08,364 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:08,364 INFO [RS:2;84486a41f81c:39799 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-04T09:43:08,364 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39799-0x101a102daf10003, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:08,364 INFO [RS:2;84486a41f81c:39799 {}] regionserver.HRegionServer(1031): Exiting; stopping=84486a41f81c,39799,1733305384625; zookeeper connection closed. 2024-12-04T09:43:08,365 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@1eea5220 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@1eea5220 2024-12-04T09:43:08,370 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/84486a41f81c,40823,1733305384489 2024-12-04T09:43:08,370 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:08,370 INFO [RS:0;84486a41f81c:40823 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-04T09:43:08,378 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [84486a41f81c,40823,1733305384489] 2024-12-04T09:43:08,386 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/84486a41f81c,40823,1733305384489 already deleted, retry=false 2024-12-04T09:43:08,386 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 84486a41f81c,40823,1733305384489 expired; onlineServers=0 2024-12-04T09:43:08,387 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '84486a41f81c,34373,1733305383783' ***** 2024-12-04T09:43:08,387 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-04T09:43:08,387 INFO [M:0;84486a41f81c:34373 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-04T09:43:08,387 INFO [M:0;84486a41f81c:34373 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-04T09:43:08,387 DEBUG [M:0;84486a41f81c:34373 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-04T09:43:08,387 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-04T09:43:08,387 DEBUG [M:0;84486a41f81c:34373 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-04T09:43:08,387 DEBUG [master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.large.0-1733305385734 {}] cleaner.HFileCleaner(306): Exit Thread[master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.large.0-1733305385734,5,FailOnTimeoutGroup] 2024-12-04T09:43:08,387 DEBUG [master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.small.0-1733305385743 {}] cleaner.HFileCleaner(306): Exit Thread[master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.small.0-1733305385743,5,FailOnTimeoutGroup] 2024-12-04T09:43:08,388 INFO [M:0;84486a41f81c:34373 {}] hbase.ChoreService(370): Chore service for: master/84486a41f81c:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-04T09:43:08,388 INFO [M:0;84486a41f81c:34373 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-04T09:43:08,388 DEBUG [M:0;84486a41f81c:34373 {}] master.HMaster(1795): Stopping service threads 2024-12-04T09:43:08,388 INFO [M:0;84486a41f81c:34373 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-04T09:43:08,388 INFO [M:0;84486a41f81c:34373 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-04T09:43:08,389 INFO [M:0;84486a41f81c:34373 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-04T09:43:08,390 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-04T09:43:08,398 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-04T09:43:08,398 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:08,398 DEBUG [M:0;84486a41f81c:34373 {}] zookeeper.ZKUtil(347): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-04T09:43:08,399 WARN [M:0;84486a41f81c:34373 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-04T09:43:08,400 INFO [M:0;84486a41f81c:34373 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/.lastflushedseqids 2024-12-04T09:43:08,412 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,412 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,414 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:50846 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:34441:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50846 dst: /127.0.0.1:34441 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:08,418 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-04T09:43:08,419 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:08,419 INFO [M:0;84486a41f81c:34373 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-04T09:43:08,419 INFO [M:0;84486a41f81c:34373 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-04T09:43:08,419 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-04T09:43:08,419 INFO [M:0;84486a41f81c:34373 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:08,419 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:08,419 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-04T09:43:08,419 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:08,420 INFO [M:0;84486a41f81c:34373 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.82 KB heapSize=34.11 KB 2024-12-04T09:43:08,437 DEBUG [M:0;84486a41f81c:34373 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/53e2ff192dbe408588693da2881e64e6 is 82, key is hbase:meta,,1/info:regioninfo/1733305386646/Put/seqid=0 2024-12-04T09:43:08,439 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,439 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,442 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:50868 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:34441:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50868 dst: /127.0.0.1:34441 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:08,445 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-04T09:43:08,446 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:08,446 INFO [M:0;84486a41f81c:34373 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/53e2ff192dbe408588693da2881e64e6 2024-12-04T09:43:08,468 DEBUG [M:0;84486a41f81c:34373 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a8858fe9c0c040c2a3e2b197002bbc18 is 747, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733305387418/Put/seqid=0 2024-12-04T09:43:08,470 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,470 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,473 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:54952 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:42541:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54952 dst: /127.0.0.1:42541 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:08,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_-9223372036854775552_1037 (size=6438) 2024-12-04T09:43:08,478 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:08,478 INFO [M:0;84486a41f81c:34373 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.13 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a8858fe9c0c040c2a3e2b197002bbc18 2024-12-04T09:43:08,478 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:08,478 INFO [RS:0;84486a41f81c:40823 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-04T09:43:08,478 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40823-0x101a102daf10001, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:08,478 INFO [RS:0;84486a41f81c:40823 {}] regionserver.HRegionServer(1031): Exiting; stopping=84486a41f81c,40823,1733305384489; zookeeper connection closed. 2024-12-04T09:43:08,479 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@3b43c82b {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@3b43c82b 2024-12-04T09:43:08,479 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-04T09:43:08,502 DEBUG [M:0;84486a41f81c:34373 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/3f7f434473374366b6959922dc53cc64 is 69, key is 84486a41f81c,34771,1733305384579/rs:state/1733305385802/Put/seqid=0 2024-12-04T09:43:08,503 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,504 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-04T09:43:08,506 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1405383319_22 at /127.0.0.1:34166 [Receiving block BP-1422175098-172.17.0.2-1733305380290:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34166 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-04T09:43:08,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-04T09:43:08,510 WARN [M:0;84486a41f81c:34373 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-04T09:43:08,510 INFO [M:0;84486a41f81c:34373 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/3f7f434473374366b6959922dc53cc64 2024-12-04T09:43:08,521 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/53e2ff192dbe408588693da2881e64e6 as hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/53e2ff192dbe408588693da2881e64e6 2024-12-04T09:43:08,529 INFO [M:0;84486a41f81c:34373 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/53e2ff192dbe408588693da2881e64e6, entries=8, sequenceid=72, filesize=5.5 K 2024-12-04T09:43:08,531 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a8858fe9c0c040c2a3e2b197002bbc18 as hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/a8858fe9c0c040c2a3e2b197002bbc18 2024-12-04T09:43:08,539 INFO [M:0;84486a41f81c:34373 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/a8858fe9c0c040c2a3e2b197002bbc18, entries=8, sequenceid=72, filesize=6.3 K 2024-12-04T09:43:08,540 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/3f7f434473374366b6959922dc53cc64 as hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/3f7f434473374366b6959922dc53cc64 2024-12-04T09:43:08,547 INFO [M:0;84486a41f81c:34373 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/3f7f434473374366b6959922dc53cc64, entries=3, sequenceid=72, filesize=5.2 K 2024-12-04T09:43:08,548 INFO [M:0;84486a41f81c:34373 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.82 KB/27459, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 129ms, sequenceid=72, compaction requested=false 2024-12-04T09:43:08,550 INFO [M:0;84486a41f81c:34373 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:08,550 DEBUG [M:0;84486a41f81c:34373 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733305388419Disabling compacts and flushes for region at 1733305388419Disabling writes for close at 1733305388419Obtaining lock to block concurrent updates at 1733305388420 (+1 ms)Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733305388420Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27459, getHeapSize=34864, getOffHeapSize=0, getCellsCount=85 at 1733305388420Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733305388421 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733305388421Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733305388436 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733305388436Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733305388453 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733305388468 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733305388468Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733305388486 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733305388501 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733305388501Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7462f76: reopening flushed file at 1733305388520 (+19 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6b7f9c74: reopening flushed file at 1733305388529 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3067fea4: reopening flushed file at 1733305388539 (+10 ms)Finished flush of dataSize ~26.82 KB/27459, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 129ms, sequenceid=72, compaction requested=false at 1733305388548 (+9 ms)Writing region close event to WAL at 1733305388550 (+2 ms)Closed at 1733305388550 2024-12-04T09:43:08,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741825_1011 (size=32662) 2024-12-04T09:43:08,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42541 is added to blk_1073741825_1011 (size=32662) 2024-12-04T09:43:08,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34441 is added to blk_1073741825_1011 (size=32662) 2024-12-04T09:43:08,554 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-04T09:43:08,554 INFO [M:0;84486a41f81c:34373 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-04T09:43:08,554 INFO [M:0;84486a41f81c:34373 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:34373 2024-12-04T09:43:08,555 INFO [M:0;84486a41f81c:34373 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-04T09:43:08,665 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:08,665 INFO [M:0;84486a41f81c:34373 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-04T09:43:08,665 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34373-0x101a102daf10000, quorum=127.0.0.1:60962, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:08,700 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2e59159d{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:08,704 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@a8e922f{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-04T09:43:08,704 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-04T09:43:08,705 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@24f92c39{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-04T09:43:08,705 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@c62369b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir/,STOPPED} 2024-12-04T09:43:08,709 WARN [BP-1422175098-172.17.0.2-1733305380290 heartbeating to localhost/127.0.0.1:37647 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-04T09:43:08,709 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-04T09:43:08,709 WARN [BP-1422175098-172.17.0.2-1733305380290 heartbeating to localhost/127.0.0.1:37647 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1422175098-172.17.0.2-1733305380290 (Datanode Uuid 6df7b36d-cc7c-4d72-b69d-8dd2ee15e86d) service to localhost/127.0.0.1:37647 2024-12-04T09:43:08,709 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-04T09:43:08,710 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data5/current/BP-1422175098-172.17.0.2-1733305380290 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:08,710 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data6/current/BP-1422175098-172.17.0.2-1733305380290 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:08,711 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-04T09:43:08,713 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1c6b8f01{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:08,713 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11f28dd2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-04T09:43:08,713 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-04T09:43:08,713 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7fa8fa5c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-04T09:43:08,713 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6463ad04{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir/,STOPPED} 2024-12-04T09:43:08,714 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-04T09:43:08,715 WARN [BP-1422175098-172.17.0.2-1733305380290 heartbeating to localhost/127.0.0.1:37647 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-04T09:43:08,715 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-04T09:43:08,715 WARN [BP-1422175098-172.17.0.2-1733305380290 heartbeating to localhost/127.0.0.1:37647 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1422175098-172.17.0.2-1733305380290 (Datanode Uuid 5f745f9f-5b17-457a-b10d-ea3589192b67) service to localhost/127.0.0.1:37647 2024-12-04T09:43:08,715 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data3/current/BP-1422175098-172.17.0.2-1733305380290 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:08,716 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data4/current/BP-1422175098-172.17.0.2-1733305380290 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:08,716 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-04T09:43:08,718 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4839957b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:08,718 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5306f615{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-04T09:43:08,718 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-04T09:43:08,718 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1a2478ad{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-04T09:43:08,718 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@550154bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir/,STOPPED} 2024-12-04T09:43:08,719 WARN [BP-1422175098-172.17.0.2-1733305380290 heartbeating to localhost/127.0.0.1:37647 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-04T09:43:08,719 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-04T09:43:08,719 WARN [BP-1422175098-172.17.0.2-1733305380290 heartbeating to localhost/127.0.0.1:37647 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1422175098-172.17.0.2-1733305380290 (Datanode Uuid 1566c7e3-5029-4d80-98d9-753760a78791) service to localhost/127.0.0.1:37647 2024-12-04T09:43:08,719 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-04T09:43:08,720 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data1/current/BP-1422175098-172.17.0.2-1733305380290 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:08,721 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/cluster_ffc2654b-e5d9-6e5a-1e51-04e5336e7a79/data/data2/current/BP-1422175098-172.17.0.2-1733305380290 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:08,721 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-04T09:43:08,732 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@76e4c45c{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-04T09:43:08,733 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-04T09:43:08,733 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-04T09:43:08,733 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-04T09:43:08,733 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir/,STOPPED} 2024-12-04T09:43:08,740 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-04T09:43:08,764 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-04T09:43:08,771 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=92 (was 161), OpenFileDescriptor=437 (was 393) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=316 (was 326), ProcessCount=11 (was 11), AvailableMemoryMB=10400 (was 10705) 2024-12-04T09:43:08,777 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=92, OpenFileDescriptor=437, MaxFileDescriptor=1048576, SystemLoadAverage=316, ProcessCount=11, AvailableMemoryMB=10400 2024-12-04T09:43:08,777 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-04T09:43:08,777 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.log.dir so I do NOT create it in target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864 2024-12-04T09:43:08,777 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f8f110f5-7000-2ce7-1118-d40a8ccf9752/hadoop.tmp.dir so I do NOT create it in target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864 2024-12-04T09:43:08,777 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8, deleteOnExit=true 2024-12-04T09:43:08,777 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-04T09:43:08,777 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/test.cache.data in system properties and HBase conf 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.tmp.dir in system properties and HBase conf 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir in system properties and HBase conf 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-04T09:43:08,778 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-04T09:43:08,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/nfs.dump.dir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/java.io.tmpdir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-04T09:43:08,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-04T09:43:09,059 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:09,064 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-04T09:43:09,065 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-04T09:43:09,065 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-04T09:43:09,065 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-04T09:43:09,066 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:09,067 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@16eaa68d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir/,AVAILABLE} 2024-12-04T09:43:09,067 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@62802e4c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-04T09:43:09,169 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3c97821d{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/java.io.tmpdir/jetty-localhost-42211-hadoop-hdfs-3_4_1-tests_jar-_-any-12993998935489681465/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-04T09:43:09,169 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@544c0dd2{HTTP/1.1, (http/1.1)}{localhost:42211} 2024-12-04T09:43:09,169 INFO [Time-limited test {}] server.Server(415): Started @10533ms 2024-12-04T09:43:09,365 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:09,369 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-04T09:43:09,373 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-04T09:43:09,373 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-04T09:43:09,373 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-04T09:43:09,374 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@61d23bc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir/,AVAILABLE} 2024-12-04T09:43:09,375 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@20a0e688{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-04T09:43:09,472 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7eeef71e{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/java.io.tmpdir/jetty-localhost-45643-hadoop-hdfs-3_4_1-tests_jar-_-any-13075225721546289634/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:09,473 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@70feba44{HTTP/1.1, (http/1.1)}{localhost:45643} 2024-12-04T09:43:09,473 INFO [Time-limited test {}] server.Server(415): Started @10836ms 2024-12-04T09:43:09,474 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-04T09:43:09,523 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:09,527 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-04T09:43:09,529 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-04T09:43:09,530 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-04T09:43:09,530 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-04T09:43:09,530 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4e5afbc4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir/,AVAILABLE} 2024-12-04T09:43:09,531 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1b0441b5{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-04T09:43:09,665 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@30bdc6f8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/java.io.tmpdir/jetty-localhost-39227-hadoop-hdfs-3_4_1-tests_jar-_-any-2000197059143249496/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:09,666 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5547eae9{HTTP/1.1, (http/1.1)}{localhost:39227} 2024-12-04T09:43:09,666 INFO [Time-limited test {}] server.Server(415): Started @11029ms 2024-12-04T09:43:09,668 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-04T09:43:09,777 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-04T09:43:09,782 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-04T09:43:09,795 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-04T09:43:09,795 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-04T09:43:09,795 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-04T09:43:09,796 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4c77de1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir/,AVAILABLE} 2024-12-04T09:43:09,796 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@438bc7ce{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-04T09:43:09,912 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@44672b71{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/java.io.tmpdir/jetty-localhost-42183-hadoop-hdfs-3_4_1-tests_jar-_-any-14655089096927879953/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:09,913 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@36a9ca95{HTTP/1.1, (http/1.1)}{localhost:42183} 2024-12-04T09:43:09,913 INFO [Time-limited test {}] server.Server(415): Started @11276ms 2024-12-04T09:43:09,917 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-04T09:43:10,152 WARN [Thread-538 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data1/current/BP-1682927199-172.17.0.2-1733305388802/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:10,171 WARN [Thread-539 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data2/current/BP-1682927199-172.17.0.2-1733305388802/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:10,199 WARN [Thread-481 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-04T09:43:10,203 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7e100d59a788d552 with lease ID 0x3cedc67acc6d4db4: Processing first storage report for DS-a2444b29-b944-43b8-b86d-c65c1648a529 from datanode DatanodeRegistration(127.0.0.1:41519, datanodeUuid=ea2e42a4-93fc-49c3-ab65-9ee31c501e86, infoPort=42291, infoSecurePort=0, ipcPort=33689, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802) 2024-12-04T09:43:10,203 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7e100d59a788d552 with lease ID 0x3cedc67acc6d4db4: from storage DS-a2444b29-b944-43b8-b86d-c65c1648a529 node DatanodeRegistration(127.0.0.1:41519, datanodeUuid=ea2e42a4-93fc-49c3-ab65-9ee31c501e86, infoPort=42291, infoSecurePort=0, ipcPort=33689, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:10,203 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7e100d59a788d552 with lease ID 0x3cedc67acc6d4db4: Processing first storage report for DS-53777a5a-49e8-46bc-8227-ef1dae490a88 from datanode DatanodeRegistration(127.0.0.1:41519, datanodeUuid=ea2e42a4-93fc-49c3-ab65-9ee31c501e86, infoPort=42291, infoSecurePort=0, ipcPort=33689, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802) 2024-12-04T09:43:10,203 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7e100d59a788d552 with lease ID 0x3cedc67acc6d4db4: from storage DS-53777a5a-49e8-46bc-8227-ef1dae490a88 node DatanodeRegistration(127.0.0.1:41519, datanodeUuid=ea2e42a4-93fc-49c3-ab65-9ee31c501e86, infoPort=42291, infoSecurePort=0, ipcPort=33689, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-04T09:43:10,452 WARN [Thread-551 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data3/current/BP-1682927199-172.17.0.2-1733305388802/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:10,455 WARN [Thread-552 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data4/current/BP-1682927199-172.17.0.2-1733305388802/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:10,484 WARN [Thread-505 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-04T09:43:10,486 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x406350d9d2bdae5e with lease ID 0x3cedc67acc6d4db5: Processing first storage report for DS-c896ebe1-175b-4716-b429-c598eaef5465 from datanode DatanodeRegistration(127.0.0.1:45145, datanodeUuid=cb5ce412-3855-4d0f-98b2-54019b9cffa1, infoPort=46657, infoSecurePort=0, ipcPort=34043, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802) 2024-12-04T09:43:10,487 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x406350d9d2bdae5e with lease ID 0x3cedc67acc6d4db5: from storage DS-c896ebe1-175b-4716-b429-c598eaef5465 node DatanodeRegistration(127.0.0.1:45145, datanodeUuid=cb5ce412-3855-4d0f-98b2-54019b9cffa1, infoPort=46657, infoSecurePort=0, ipcPort=34043, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:10,487 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x406350d9d2bdae5e with lease ID 0x3cedc67acc6d4db5: Processing first storage report for DS-800ab53b-b477-45a2-9987-5ba2c7e925d9 from datanode DatanodeRegistration(127.0.0.1:45145, datanodeUuid=cb5ce412-3855-4d0f-98b2-54019b9cffa1, infoPort=46657, infoSecurePort=0, ipcPort=34043, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802) 2024-12-04T09:43:10,487 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x406350d9d2bdae5e with lease ID 0x3cedc67acc6d4db5: from storage DS-800ab53b-b477-45a2-9987-5ba2c7e925d9 node DatanodeRegistration(127.0.0.1:45145, datanodeUuid=cb5ce412-3855-4d0f-98b2-54019b9cffa1, infoPort=46657, infoSecurePort=0, ipcPort=34043, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:10,650 WARN [Thread-564 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data6/current/BP-1682927199-172.17.0.2-1733305388802/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:10,650 WARN [Thread-563 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data5/current/BP-1682927199-172.17.0.2-1733305388802/current, will proceed with Du for space computation calculation, 2024-12-04T09:43:10,684 WARN [Thread-528 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-04T09:43:10,687 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5e6879f2b4d2553 with lease ID 0x3cedc67acc6d4db6: Processing first storage report for DS-4fb2253d-851b-45ef-9f13-2c5552550cda from datanode DatanodeRegistration(127.0.0.1:43879, datanodeUuid=efc6e4c9-c5aa-4888-9f32-4a1d33a8ec16, infoPort=35797, infoSecurePort=0, ipcPort=38401, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802) 2024-12-04T09:43:10,688 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5e6879f2b4d2553 with lease ID 0x3cedc67acc6d4db6: from storage DS-4fb2253d-851b-45ef-9f13-2c5552550cda node DatanodeRegistration(127.0.0.1:43879, datanodeUuid=efc6e4c9-c5aa-4888-9f32-4a1d33a8ec16, infoPort=35797, infoSecurePort=0, ipcPort=38401, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:10,688 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5e6879f2b4d2553 with lease ID 0x3cedc67acc6d4db6: Processing first storage report for DS-8e28a0e0-cf92-43d3-b2a6-8a4a3acea486 from datanode DatanodeRegistration(127.0.0.1:43879, datanodeUuid=efc6e4c9-c5aa-4888-9f32-4a1d33a8ec16, infoPort=35797, infoSecurePort=0, ipcPort=38401, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802) 2024-12-04T09:43:10,688 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5e6879f2b4d2553 with lease ID 0x3cedc67acc6d4db6: from storage DS-8e28a0e0-cf92-43d3-b2a6-8a4a3acea486 node DatanodeRegistration(127.0.0.1:43879, datanodeUuid=efc6e4c9-c5aa-4888-9f32-4a1d33a8ec16, infoPort=35797, infoSecurePort=0, ipcPort=38401, storageInfo=lv=-57;cid=testClusterID;nsid=2047110580;c=1733305388802), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-04T09:43:10,776 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864 2024-12-04T09:43:10,778 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/zookeeper_0, clientPort=51778, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-04T09:43:10,779 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=51778 2024-12-04T09:43:10,780 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:10,782 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:10,795 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741825_1001 (size=7) 2024-12-04T09:43:10,795 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741825_1001 (size=7) 2024-12-04T09:43:10,796 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741825_1001 (size=7) 2024-12-04T09:43:10,797 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0 with version=8 2024-12-04T09:43:10,797 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:37647/user/jenkins/test-data/6fa98056-54bf-0edd-a2dc-3478aa8b9241/hbase-staging 2024-12-04T09:43:10,800 INFO [Time-limited test {}] client.ConnectionUtils(128): master/84486a41f81c:0 server-side Connection retries=45 2024-12-04T09:43:10,800 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,800 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,800 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-04T09:43:10,800 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,800 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-04T09:43:10,800 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-04T09:43:10,800 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-04T09:43:10,801 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:33655 2024-12-04T09:43:10,802 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:33655 connecting to ZooKeeper ensemble=127.0.0.1:51778 2024-12-04T09:43:10,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:336550x0, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-04T09:43:10,852 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:33655-0x101a102f92f0000 connected 2024-12-04T09:43:10,923 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:10,925 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:10,927 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:10,927 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0, hbase.cluster.distributed=false 2024-12-04T09:43:10,929 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-04T09:43:10,930 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=33655 2024-12-04T09:43:10,933 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=33655 2024-12-04T09:43:10,934 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=33655 2024-12-04T09:43:10,934 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=33655 2024-12-04T09:43:10,934 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=33655 2024-12-04T09:43:10,950 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/84486a41f81c:0 server-side Connection retries=45 2024-12-04T09:43:10,950 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,950 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,950 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-04T09:43:10,950 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,951 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-04T09:43:10,951 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-04T09:43:10,951 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-04T09:43:10,952 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:40111 2024-12-04T09:43:10,953 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:40111 connecting to ZooKeeper ensemble=127.0.0.1:51778 2024-12-04T09:43:10,955 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:10,956 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:10,970 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:401110x0, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-04T09:43:10,970 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:401110x0, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:10,970 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:40111-0x101a102f92f0001 connected 2024-12-04T09:43:10,971 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-04T09:43:10,971 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-04T09:43:10,972 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-04T09:43:10,974 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-04T09:43:10,974 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40111 2024-12-04T09:43:10,975 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40111 2024-12-04T09:43:10,975 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40111 2024-12-04T09:43:10,977 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40111 2024-12-04T09:43:10,977 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40111 2024-12-04T09:43:10,992 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/84486a41f81c:0 server-side Connection retries=45 2024-12-04T09:43:10,992 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,992 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,992 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-04T09:43:10,992 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:10,992 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-04T09:43:10,992 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-04T09:43:10,992 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-04T09:43:10,993 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39675 2024-12-04T09:43:10,994 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39675 connecting to ZooKeeper ensemble=127.0.0.1:51778 2024-12-04T09:43:10,995 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:10,997 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:11,012 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:396750x0, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-04T09:43:11,012 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39675-0x101a102f92f0002 connected 2024-12-04T09:43:11,012 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:11,013 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-04T09:43:11,015 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-04T09:43:11,016 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-04T09:43:11,017 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-04T09:43:11,018 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39675 2024-12-04T09:43:11,019 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39675 2024-12-04T09:43:11,021 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39675 2024-12-04T09:43:11,021 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39675 2024-12-04T09:43:11,022 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39675 2024-12-04T09:43:11,037 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/84486a41f81c:0 server-side Connection retries=45 2024-12-04T09:43:11,038 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:11,038 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:11,038 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-04T09:43:11,038 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-04T09:43:11,038 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-04T09:43:11,038 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-04T09:43:11,038 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-04T09:43:11,039 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:32869 2024-12-04T09:43:11,040 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:32869 connecting to ZooKeeper ensemble=127.0.0.1:51778 2024-12-04T09:43:11,041 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:11,043 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:11,053 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:328690x0, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-04T09:43:11,054 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:328690x0, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:11,054 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:32869-0x101a102f92f0003 connected 2024-12-04T09:43:11,054 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-04T09:43:11,056 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-04T09:43:11,057 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-04T09:43:11,059 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-04T09:43:11,062 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=32869 2024-12-04T09:43:11,062 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=32869 2024-12-04T09:43:11,062 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=32869 2024-12-04T09:43:11,062 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=32869 2024-12-04T09:43:11,063 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=32869 2024-12-04T09:43:11,079 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;84486a41f81c:33655 2024-12-04T09:43:11,080 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/84486a41f81c,33655,1733305390799 2024-12-04T09:43:11,089 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:11,089 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:11,089 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:11,089 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:11,090 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/84486a41f81c,33655,1733305390799 2024-12-04T09:43:11,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-04T09:43:11,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-04T09:43:11,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,103 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-04T09:43:11,103 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,104 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-04T09:43:11,105 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/84486a41f81c,33655,1733305390799 from backup master directory 2024-12-04T09:43:11,114 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:11,114 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:11,114 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/84486a41f81c,33655,1733305390799 2024-12-04T09:43:11,114 WARN [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-04T09:43:11,114 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:11,115 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=84486a41f81c,33655,1733305390799 2024-12-04T09:43:11,115 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-04T09:43:11,121 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/hbase.id] with ID: b3986b62-462d-4435-a67e-2edc067e4ddb 2024-12-04T09:43:11,121 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/.tmp/hbase.id 2024-12-04T09:43:11,134 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741826_1002 (size=42) 2024-12-04T09:43:11,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741826_1002 (size=42) 2024-12-04T09:43:11,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741826_1002 (size=42) 2024-12-04T09:43:11,136 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/.tmp/hbase.id]:[hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/hbase.id] 2024-12-04T09:43:11,155 INFO [master/84486a41f81c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-04T09:43:11,155 INFO [master/84486a41f81c:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-04T09:43:11,157 INFO [master/84486a41f81c:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-12-04T09:43:11,164 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,164 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,164 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,167 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741827_1003 (size=196) 2024-12-04T09:43:11,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741827_1003 (size=196) 2024-12-04T09:43:11,180 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-04T09:43:11,181 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-04T09:43:11,181 INFO [master/84486a41f81c:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-04T09:43:11,182 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741827_1003 (size=196) 2024-12-04T09:43:11,194 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741828_1004 (size=1189) 2024-12-04T09:43:11,194 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741828_1004 (size=1189) 2024-12-04T09:43:11,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741828_1004 (size=1189) 2024-12-04T09:43:11,196 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store 2024-12-04T09:43:11,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741829_1005 (size=34) 2024-12-04T09:43:11,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741829_1005 (size=34) 2024-12-04T09:43:11,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741829_1005 (size=34) 2024-12-04T09:43:11,207 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:11,207 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-04T09:43:11,207 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:11,207 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:11,207 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-04T09:43:11,208 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:11,208 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:11,208 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733305391207Disabling compacts and flushes for region at 1733305391207Disabling writes for close at 1733305391208 (+1 ms)Writing region close event to WAL at 1733305391208Closed at 1733305391208 2024-12-04T09:43:11,209 WARN [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/.initializing 2024-12-04T09:43:11,209 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/WALs/84486a41f81c,33655,1733305390799 2024-12-04T09:43:11,214 INFO [master/84486a41f81c:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C33655%2C1733305390799, suffix=, logDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/WALs/84486a41f81c,33655,1733305390799, archiveDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/oldWALs, maxLogs=10 2024-12-04T09:43:11,215 INFO [master/84486a41f81c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 84486a41f81c%2C33655%2C1733305390799.1733305391214 2024-12-04T09:43:11,229 INFO [master/84486a41f81c:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/WALs/84486a41f81c,33655,1733305390799/84486a41f81c%2C33655%2C1733305390799.1733305391214 2024-12-04T09:43:11,234 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35797:35797),(127.0.0.1/127.0.0.1:46657:46657),(127.0.0.1/127.0.0.1:42291:42291)] 2024-12-04T09:43:11,240 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-04T09:43:11,240 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:11,240 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,240 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,244 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,246 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-04T09:43:11,247 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,248 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:11,248 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,250 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-04T09:43:11,250 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,251 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-04T09:43:11,252 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,254 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-04T09:43:11,255 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,256 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-04T09:43:11,256 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,258 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-04T09:43:11,258 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,259 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-04T09:43:11,259 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,260 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,260 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,262 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,262 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,262 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-04T09:43:11,264 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-04T09:43:11,266 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-04T09:43:11,267 INFO [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60260031, jitterRate=-0.10205556452274323}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-04T09:43:11,268 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733305391241Initializing all the Stores at 1733305391243 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305391243Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305391244 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305391244Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305391244Cleaning up temporary data from old regions at 1733305391262 (+18 ms)Region opened successfully at 1733305391268 (+6 ms) 2024-12-04T09:43:11,268 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-04T09:43:11,272 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@18986ed4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=84486a41f81c/172.17.0.2:0 2024-12-04T09:43:11,273 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-04T09:43:11,273 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-04T09:43:11,273 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-04T09:43:11,274 INFO [master/84486a41f81c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-04T09:43:11,274 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-04T09:43:11,275 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-04T09:43:11,275 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-04T09:43:11,278 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-04T09:43:11,279 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-04T09:43:11,289 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-04T09:43:11,290 INFO [master/84486a41f81c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-04T09:43:11,292 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-04T09:43:11,306 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-04T09:43:11,307 INFO [master/84486a41f81c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-04T09:43:11,308 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-04T09:43:11,314 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-04T09:43:11,316 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-04T09:43:11,323 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-04T09:43:11,326 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-04T09:43:11,336 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-04T09:43:11,345 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:11,345 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:11,345 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:11,345 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,345 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:11,345 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,345 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,346 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,346 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=84486a41f81c,33655,1733305390799, sessionid=0x101a102f92f0000, setting cluster-up flag (Was=false) 2024-12-04T09:43:11,364 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,364 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,364 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,364 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,389 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-04T09:43:11,390 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=84486a41f81c,33655,1733305390799 2024-12-04T09:43:11,406 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,406 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,406 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,406 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:11,431 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-04T09:43:11,435 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=84486a41f81c,33655,1733305390799 2024-12-04T09:43:11,437 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-04T09:43:11,440 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-04T09:43:11,440 INFO [master/84486a41f81c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-04T09:43:11,440 INFO [master/84486a41f81c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-04T09:43:11,441 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 84486a41f81c,33655,1733305390799 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-04T09:43:11,442 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/84486a41f81c:0, corePoolSize=5, maxPoolSize=5 2024-12-04T09:43:11,443 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/84486a41f81c:0, corePoolSize=5, maxPoolSize=5 2024-12-04T09:43:11,443 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/84486a41f81c:0, corePoolSize=5, maxPoolSize=5 2024-12-04T09:43:11,443 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/84486a41f81c:0, corePoolSize=5, maxPoolSize=5 2024-12-04T09:43:11,443 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/84486a41f81c:0, corePoolSize=10, maxPoolSize=10 2024-12-04T09:43:11,443 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,443 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/84486a41f81c:0, corePoolSize=2, maxPoolSize=2 2024-12-04T09:43:11,443 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,448 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733305421448 2024-12-04T09:43:11,449 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-04T09:43:11,449 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-04T09:43:11,449 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-04T09:43:11,449 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-04T09:43:11,449 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-04T09:43:11,449 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-04T09:43:11,449 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-04T09:43:11,449 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-04T09:43:11,449 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,450 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-04T09:43:11,451 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-04T09:43:11,451 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-04T09:43:11,451 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-04T09:43:11,451 INFO [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-04T09:43:11,451 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.large.0-1733305391451,5,FailOnTimeoutGroup] 2024-12-04T09:43:11,451 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.small.0-1733305391451,5,FailOnTimeoutGroup] 2024-12-04T09:43:11,452 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,452 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,452 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-04T09:43:11,452 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,452 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,452 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-04T09:43:11,468 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(746): ClusterId : b3986b62-462d-4435-a67e-2edc067e4ddb 2024-12-04T09:43:11,468 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(746): ClusterId : b3986b62-462d-4435-a67e-2edc067e4ddb 2024-12-04T09:43:11,469 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-04T09:43:11,469 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-04T09:43:11,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741831_1007 (size=1321) 2024-12-04T09:43:11,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741831_1007 (size=1321) 2024-12-04T09:43:11,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741831_1007 (size=1321) 2024-12-04T09:43:11,478 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(746): ClusterId : b3986b62-462d-4435-a67e-2edc067e4ddb 2024-12-04T09:43:11,478 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-04T09:43:11,480 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-04T09:43:11,480 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0 2024-12-04T09:43:11,495 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741832_1008 (size=32) 2024-12-04T09:43:11,495 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741832_1008 (size=32) 2024-12-04T09:43:11,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741832_1008 (size=32) 2024-12-04T09:43:11,501 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:11,502 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-04T09:43:11,502 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-04T09:43:11,504 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-04T09:43:11,505 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-04T09:43:11,505 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-04T09:43:11,506 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-04T09:43:11,506 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,507 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:11,507 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-04T09:43:11,509 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-04T09:43:11,509 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,509 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:11,510 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-04T09:43:11,512 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-04T09:43:11,512 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,513 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:11,513 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-04T09:43:11,520 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-04T09:43:11,520 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:11,521 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:11,521 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-04T09:43:11,522 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740 2024-12-04T09:43:11,522 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740 2024-12-04T09:43:11,523 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-04T09:43:11,523 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-04T09:43:11,524 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-04T09:43:11,524 DEBUG [RS:0;84486a41f81c:40111 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5a8b3f6b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=84486a41f81c/172.17.0.2:0 2024-12-04T09:43:11,524 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-04T09:43:11,524 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-04T09:43:11,525 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-04T09:43:11,527 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-04T09:43:11,535 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-04T09:43:11,536 DEBUG [RS:0;84486a41f81c:40111 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;84486a41f81c:40111 2024-12-04T09:43:11,536 INFO [RS:0;84486a41f81c:40111 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-04T09:43:11,536 INFO [RS:0;84486a41f81c:40111 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-04T09:43:11,536 DEBUG [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-04T09:43:11,536 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63678361, jitterRate=-0.051118478178977966}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-04T09:43:11,536 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-04T09:43:11,537 DEBUG [RS:1;84486a41f81c:39675 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@67e3a2aa, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=84486a41f81c/172.17.0.2:0 2024-12-04T09:43:11,537 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733305391501Initializing all the Stores at 1733305391503 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305391503Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305391503Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305391503Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305391503Cleaning up temporary data from old regions at 1733305391524 (+21 ms)Region opened successfully at 1733305391537 (+13 ms) 2024-12-04T09:43:11,537 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(2659): reportForDuty to master=84486a41f81c,33655,1733305390799 with port=40111, startcode=1733305390950 2024-12-04T09:43:11,537 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-04T09:43:11,537 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-04T09:43:11,537 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-04T09:43:11,537 DEBUG [RS:0;84486a41f81c:40111 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-04T09:43:11,537 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-04T09:43:11,537 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-04T09:43:11,537 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-04T09:43:11,538 DEBUG [RS:2;84486a41f81c:32869 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1a5622ed, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=84486a41f81c/172.17.0.2:0 2024-12-04T09:43:11,539 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-04T09:43:11,539 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733305391537Disabling compacts and flushes for region at 1733305391537Disabling writes for close at 1733305391537Writing region close event to WAL at 1733305391539 (+2 ms)Closed at 1733305391539 2024-12-04T09:43:11,541 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56609, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-04T09:43:11,542 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33655 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 84486a41f81c,40111,1733305390950 2024-12-04T09:43:11,542 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33655 {}] master.ServerManager(517): Registering regionserver=84486a41f81c,40111,1733305390950 2024-12-04T09:43:11,547 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-04T09:43:11,547 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-04T09:43:11,547 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-04T09:43:11,549 DEBUG [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0 2024-12-04T09:43:11,549 DEBUG [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:34577 2024-12-04T09:43:11,549 DEBUG [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-04T09:43:11,550 DEBUG [RS:1;84486a41f81c:39675 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;84486a41f81c:39675 2024-12-04T09:43:11,550 INFO [RS:1;84486a41f81c:39675 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-04T09:43:11,550 INFO [RS:1;84486a41f81c:39675 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-04T09:43:11,550 DEBUG [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-04T09:43:11,551 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-04T09:43:11,553 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-04T09:43:11,553 DEBUG [RS:2;84486a41f81c:32869 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;84486a41f81c:32869 2024-12-04T09:43:11,553 INFO [RS:2;84486a41f81c:32869 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-04T09:43:11,553 INFO [RS:2;84486a41f81c:32869 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-04T09:43:11,553 DEBUG [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-04T09:43:11,556 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:11,557 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(2659): reportForDuty to master=84486a41f81c,33655,1733305390799 with port=39675, startcode=1733305390992 2024-12-04T09:43:11,557 DEBUG [RS:1;84486a41f81c:39675 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-04T09:43:11,557 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [84486a41f81c,40111,1733305390950] 2024-12-04T09:43:11,558 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(2659): reportForDuty to master=84486a41f81c,33655,1733305390799 with port=32869, startcode=1733305391037 2024-12-04T09:43:11,559 DEBUG [RS:0;84486a41f81c:40111 {}] zookeeper.ZKUtil(111): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/84486a41f81c,40111,1733305390950 2024-12-04T09:43:11,559 WARN [RS:0;84486a41f81c:40111 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-04T09:43:11,559 DEBUG [RS:2;84486a41f81c:32869 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-04T09:43:11,559 INFO [RS:0;84486a41f81c:40111 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-04T09:43:11,559 DEBUG [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,40111,1733305390950 2024-12-04T09:43:11,559 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42769, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-04T09:43:11,559 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33655 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 84486a41f81c,39675,1733305390992 2024-12-04T09:43:11,560 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33655 {}] master.ServerManager(517): Registering regionserver=84486a41f81c,39675,1733305390992 2024-12-04T09:43:11,560 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58979, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-04T09:43:11,562 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33655 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 84486a41f81c,32869,1733305391037 2024-12-04T09:43:11,562 DEBUG [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0 2024-12-04T09:43:11,562 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33655 {}] master.ServerManager(517): Registering regionserver=84486a41f81c,32869,1733305391037 2024-12-04T09:43:11,562 DEBUG [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:34577 2024-12-04T09:43:11,562 DEBUG [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-04T09:43:11,564 DEBUG [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0 2024-12-04T09:43:11,564 DEBUG [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:34577 2024-12-04T09:43:11,564 DEBUG [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-04T09:43:11,591 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:11,606 DEBUG [RS:1;84486a41f81c:39675 {}] zookeeper.ZKUtil(111): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/84486a41f81c,39675,1733305390992 2024-12-04T09:43:11,606 WARN [RS:1;84486a41f81c:39675 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-04T09:43:11,606 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [84486a41f81c,32869,1733305391037] 2024-12-04T09:43:11,606 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [84486a41f81c,39675,1733305390992] 2024-12-04T09:43:11,606 INFO [RS:1;84486a41f81c:39675 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-04T09:43:11,606 DEBUG [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,39675,1733305390992 2024-12-04T09:43:11,606 DEBUG [RS:2;84486a41f81c:32869 {}] zookeeper.ZKUtil(111): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/84486a41f81c,32869,1733305391037 2024-12-04T09:43:11,607 WARN [RS:2;84486a41f81c:32869 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-04T09:43:11,607 INFO [RS:2;84486a41f81c:32869 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-04T09:43:11,607 DEBUG [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,32869,1733305391037 2024-12-04T09:43:11,607 INFO [RS:0;84486a41f81c:40111 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-04T09:43:11,609 INFO [RS:0;84486a41f81c:40111 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-04T09:43:11,609 INFO [RS:0;84486a41f81c:40111 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-04T09:43:11,609 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,611 INFO [RS:1;84486a41f81c:39675 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-04T09:43:11,616 INFO [RS:2;84486a41f81c:32869 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-04T09:43:11,618 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-04T09:43:11,618 INFO [RS:1;84486a41f81c:39675 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-04T09:43:11,618 INFO [RS:2;84486a41f81c:32869 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-04T09:43:11,619 INFO [RS:0;84486a41f81c:40111 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-04T09:43:11,619 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,620 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,620 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,620 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,620 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,620 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,620 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/84486a41f81c:0, corePoolSize=2, maxPoolSize=2 2024-12-04T09:43:11,620 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,621 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,621 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,621 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,621 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,621 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,621 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:11,621 DEBUG [RS:0;84486a41f81c:40111 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:11,624 INFO [RS:2;84486a41f81c:32869 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-04T09:43:11,624 INFO [RS:1;84486a41f81c:39675 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-04T09:43:11,624 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,624 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,625 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-04T09:43:11,626 INFO [RS:2;84486a41f81c:32869 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-04T09:43:11,626 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,626 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,626 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,626 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-04T09:43:11,626 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,626 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,626 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,626 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/84486a41f81c:0, corePoolSize=2, maxPoolSize=2 2024-12-04T09:43:11,626 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:11,627 DEBUG [RS:2;84486a41f81c:32869 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:11,627 INFO [RS:1;84486a41f81c:39675 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-04T09:43:11,627 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,627 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,627 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/84486a41f81c:0, corePoolSize=2, maxPoolSize=2 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/84486a41f81c:0, corePoolSize=1, maxPoolSize=1 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:11,628 DEBUG [RS:1;84486a41f81c:39675 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0, corePoolSize=3, maxPoolSize=3 2024-12-04T09:43:11,629 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,629 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,629 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,629 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,629 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,629 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,40111,1733305390950-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-04T09:43:11,629 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,32869,1733305391037-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-04T09:43:11,630 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,39675,1733305390992-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-04T09:43:11,643 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-04T09:43:11,643 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,40111,1733305390950-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,643 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-04T09:43:11,643 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,644 INFO [RS:0;84486a41f81c:40111 {}] regionserver.Replication(171): 84486a41f81c,40111,1733305390950 started 2024-12-04T09:43:11,644 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,39675,1733305390992-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,644 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,644 INFO [RS:1;84486a41f81c:39675 {}] regionserver.Replication(171): 84486a41f81c,39675,1733305390992 started 2024-12-04T09:43:11,647 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-04T09:43:11,648 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,32869,1733305391037-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,648 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,648 INFO [RS:2;84486a41f81c:32869 {}] regionserver.Replication(171): 84486a41f81c,32869,1733305391037 started 2024-12-04T09:43:11,659 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,659 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1482): Serving as 84486a41f81c,40111,1733305390950, RpcServer on 84486a41f81c/172.17.0.2:40111, sessionid=0x101a102f92f0001 2024-12-04T09:43:11,659 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-04T09:43:11,659 DEBUG [RS:0;84486a41f81c:40111 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 84486a41f81c,40111,1733305390950 2024-12-04T09:43:11,659 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,40111,1733305390950' 2024-12-04T09:43:11,659 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-04T09:43:11,660 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-04T09:43:11,660 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-04T09:43:11,660 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-04T09:43:11,660 DEBUG [RS:0;84486a41f81c:40111 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 84486a41f81c,40111,1733305390950 2024-12-04T09:43:11,660 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,40111,1733305390950' 2024-12-04T09:43:11,660 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-04T09:43:11,661 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-04T09:43:11,661 DEBUG [RS:0;84486a41f81c:40111 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-04T09:43:11,661 INFO [RS:0;84486a41f81c:40111 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-04T09:43:11,661 INFO [RS:0;84486a41f81c:40111 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-04T09:43:11,663 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,663 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:11,664 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(1482): Serving as 84486a41f81c,39675,1733305390992, RpcServer on 84486a41f81c/172.17.0.2:39675, sessionid=0x101a102f92f0002 2024-12-04T09:43:11,664 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1482): Serving as 84486a41f81c,32869,1733305391037, RpcServer on 84486a41f81c/172.17.0.2:32869, sessionid=0x101a102f92f0003 2024-12-04T09:43:11,664 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-04T09:43:11,664 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-04T09:43:11,664 DEBUG [RS:1;84486a41f81c:39675 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 84486a41f81c,39675,1733305390992 2024-12-04T09:43:11,664 DEBUG [RS:2;84486a41f81c:32869 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 84486a41f81c,32869,1733305391037 2024-12-04T09:43:11,664 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,39675,1733305390992' 2024-12-04T09:43:11,664 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,32869,1733305391037' 2024-12-04T09:43:11,664 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-04T09:43:11,664 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-04T09:43:11,665 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-04T09:43:11,665 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-04T09:43:11,665 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-04T09:43:11,665 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-04T09:43:11,665 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-04T09:43:11,665 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-04T09:43:11,665 DEBUG [RS:1;84486a41f81c:39675 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 84486a41f81c,39675,1733305390992 2024-12-04T09:43:11,665 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,39675,1733305390992' 2024-12-04T09:43:11,665 DEBUG [RS:2;84486a41f81c:32869 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 84486a41f81c,32869,1733305391037 2024-12-04T09:43:11,665 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-04T09:43:11,665 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '84486a41f81c,32869,1733305391037' 2024-12-04T09:43:11,665 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-04T09:43:11,666 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-04T09:43:11,666 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-04T09:43:11,666 DEBUG [RS:1;84486a41f81c:39675 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-04T09:43:11,666 INFO [RS:1;84486a41f81c:39675 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-04T09:43:11,666 INFO [RS:1;84486a41f81c:39675 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-04T09:43:11,666 DEBUG [RS:2;84486a41f81c:32869 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-04T09:43:11,666 INFO [RS:2;84486a41f81c:32869 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-04T09:43:11,666 INFO [RS:2;84486a41f81c:32869 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-04T09:43:11,703 WARN [84486a41f81c:33655 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-04T09:43:11,764 INFO [RS:0;84486a41f81c:40111 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C40111%2C1733305390950, suffix=, logDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,40111,1733305390950, archiveDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/oldWALs, maxLogs=32 2024-12-04T09:43:11,766 INFO [RS:0;84486a41f81c:40111 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 84486a41f81c%2C40111%2C1733305390950.1733305391766 2024-12-04T09:43:11,769 INFO [RS:1;84486a41f81c:39675 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C39675%2C1733305390992, suffix=, logDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,39675,1733305390992, archiveDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/oldWALs, maxLogs=32 2024-12-04T09:43:11,771 INFO [RS:1;84486a41f81c:39675 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 84486a41f81c%2C39675%2C1733305390992.1733305391770 2024-12-04T09:43:11,771 INFO [RS:2;84486a41f81c:32869 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C32869%2C1733305391037, suffix=, logDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,32869,1733305391037, archiveDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/oldWALs, maxLogs=32 2024-12-04T09:43:11,773 INFO [RS:2;84486a41f81c:32869 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 84486a41f81c%2C32869%2C1733305391037.1733305391773 2024-12-04T09:43:11,781 INFO [RS:0;84486a41f81c:40111 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,40111,1733305390950/84486a41f81c%2C40111%2C1733305390950.1733305391766 2024-12-04T09:43:11,786 INFO [RS:1;84486a41f81c:39675 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,39675,1733305390992/84486a41f81c%2C39675%2C1733305390992.1733305391770 2024-12-04T09:43:11,790 DEBUG [RS:0;84486a41f81c:40111 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46657:46657),(127.0.0.1/127.0.0.1:35797:35797),(127.0.0.1/127.0.0.1:42291:42291)] 2024-12-04T09:43:11,790 DEBUG [RS:1;84486a41f81c:39675 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46657:46657),(127.0.0.1/127.0.0.1:42291:42291),(127.0.0.1/127.0.0.1:35797:35797)] 2024-12-04T09:43:11,791 INFO [RS:2;84486a41f81c:32869 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,32869,1733305391037/84486a41f81c%2C32869%2C1733305391037.1733305391773 2024-12-04T09:43:11,798 DEBUG [RS:2;84486a41f81c:32869 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:42291:42291),(127.0.0.1/127.0.0.1:46657:46657),(127.0.0.1/127.0.0.1:35797:35797)] 2024-12-04T09:43:11,953 DEBUG [84486a41f81c:33655 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-04T09:43:11,954 DEBUG [84486a41f81c:33655 {}] balancer.BalancerClusterState(204): Hosts are {84486a41f81c=0} racks are {/default-rack=0} 2024-12-04T09:43:11,956 DEBUG [84486a41f81c:33655 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-04T09:43:11,956 DEBUG [84486a41f81c:33655 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-04T09:43:11,956 DEBUG [84486a41f81c:33655 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-04T09:43:11,956 DEBUG [84486a41f81c:33655 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-04T09:43:11,956 DEBUG [84486a41f81c:33655 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-04T09:43:11,956 DEBUG [84486a41f81c:33655 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-04T09:43:11,956 INFO [84486a41f81c:33655 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-04T09:43:11,956 INFO [84486a41f81c:33655 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-04T09:43:11,956 INFO [84486a41f81c:33655 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-04T09:43:11,956 DEBUG [84486a41f81c:33655 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-04T09:43:11,957 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=84486a41f81c,32869,1733305391037 2024-12-04T09:43:11,959 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 84486a41f81c,32869,1733305391037, state=OPENING 2024-12-04T09:43:12,105 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-04T09:43:12,115 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-04T09:43:12,163 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-04T09:43:12,163 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-04T09:43:12,275 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-04T09:43:12,313 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:12,313 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:12,313 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:12,313 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:12,314 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-04T09:43:12,314 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:12,314 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:12,314 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=84486a41f81c,32869,1733305391037}] 2024-12-04T09:43:12,314 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:12,315 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:12,468 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-04T09:43:12,470 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36261, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-04T09:43:12,475 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-04T09:43:12,475 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-04T09:43:12,477 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=84486a41f81c%2C32869%2C1733305391037.meta, suffix=.meta, logDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,32869,1733305391037, archiveDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/oldWALs, maxLogs=32 2024-12-04T09:43:12,479 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 84486a41f81c%2C32869%2C1733305391037.meta.1733305392478.meta 2024-12-04T09:43:12,494 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/WALs/84486a41f81c,32869,1733305391037/84486a41f81c%2C32869%2C1733305391037.meta.1733305392478.meta 2024-12-04T09:43:12,496 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:42291:42291),(127.0.0.1/127.0.0.1:35797:35797),(127.0.0.1/127.0.0.1:46657:46657)] 2024-12-04T09:43:12,498 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-04T09:43:12,498 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-04T09:43:12,498 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-04T09:43:12,498 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-04T09:43:12,498 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-04T09:43:12,499 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:12,499 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-04T09:43:12,499 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-04T09:43:12,500 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-04T09:43:12,501 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-04T09:43:12,501 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:12,502 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:12,502 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-04T09:43:12,503 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-04T09:43:12,503 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:12,503 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:12,503 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-04T09:43:12,504 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-04T09:43:12,504 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:12,505 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:12,505 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-04T09:43:12,506 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-04T09:43:12,506 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:12,506 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-04T09:43:12,506 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-04T09:43:12,507 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740 2024-12-04T09:43:12,508 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740 2024-12-04T09:43:12,509 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-04T09:43:12,509 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-04T09:43:12,510 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-04T09:43:12,512 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-04T09:43:12,513 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71771650, jitterRate=0.06948092579841614}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-04T09:43:12,513 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-04T09:43:12,515 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733305392499Writing region info on filesystem at 1733305392499Initializing all the Stores at 1733305392500 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305392500Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305392500Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305392500Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733305392500Cleaning up temporary data from old regions at 1733305392509 (+9 ms)Running coprocessor post-open hooks at 1733305392513 (+4 ms)Region opened successfully at 1733305392515 (+2 ms) 2024-12-04T09:43:12,516 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733305392468 2024-12-04T09:43:12,520 DEBUG [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-04T09:43:12,520 INFO [RS_OPEN_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-04T09:43:12,521 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=84486a41f81c,32869,1733305391037 2024-12-04T09:43:12,522 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 84486a41f81c,32869,1733305391037, state=OPEN 2024-12-04T09:43:12,536 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-04T09:43:12,536 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-04T09:43:12,536 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-04T09:43:12,536 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-04T09:43:12,537 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:12,537 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:12,537 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:12,537 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-04T09:43:12,537 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=84486a41f81c,32869,1733305391037 2024-12-04T09:43:12,542 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-04T09:43:12,542 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=84486a41f81c,32869,1733305391037 in 223 msec 2024-12-04T09:43:12,546 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-04T09:43:12,546 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 995 msec 2024-12-04T09:43:12,547 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-04T09:43:12,547 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-04T09:43:12,548 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-04T09:43:12,549 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=84486a41f81c,32869,1733305391037, seqNum=-1] 2024-12-04T09:43:12,549 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-04T09:43:12,551 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56443, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-04T09:43:12,559 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1180 sec 2024-12-04T09:43:12,559 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733305392559, completionTime=-1 2024-12-04T09:43:12,559 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-04T09:43:12,559 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-04T09:43:12,562 INFO [master/84486a41f81c:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-04T09:43:12,562 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733305452562 2024-12-04T09:43:12,562 INFO [master/84486a41f81c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733305512562 2024-12-04T09:43:12,562 INFO [master/84486a41f81c:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-12-04T09:43:12,562 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,33655,1733305390799-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:12,562 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,33655,1733305390799-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:12,562 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,33655,1733305390799-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:12,563 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-84486a41f81c:33655, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:12,563 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:12,563 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:12,565 DEBUG [master/84486a41f81c:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-04T09:43:12,570 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.455sec 2024-12-04T09:43:12,571 INFO [master/84486a41f81c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-04T09:43:12,571 INFO [master/84486a41f81c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-04T09:43:12,571 INFO [master/84486a41f81c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-04T09:43:12,571 INFO [master/84486a41f81c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-04T09:43:12,571 INFO [master/84486a41f81c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-04T09:43:12,571 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,33655,1733305390799-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-04T09:43:12,571 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,33655,1733305390799-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-04T09:43:12,574 DEBUG [master/84486a41f81c:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-04T09:43:12,574 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-04T09:43:12,574 INFO [master/84486a41f81c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=84486a41f81c,33655,1733305390799-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-04T09:43:12,669 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@10f5e381, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-04T09:43:12,669 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 84486a41f81c,33655,-1 for getting cluster id 2024-12-04T09:43:12,670 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-04T09:43:12,671 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'b3986b62-462d-4435-a67e-2edc067e4ddb' 2024-12-04T09:43:12,671 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-04T09:43:12,672 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "b3986b62-462d-4435-a67e-2edc067e4ddb" 2024-12-04T09:43:12,672 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@59093b2c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-04T09:43:12,672 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [84486a41f81c,33655,-1] 2024-12-04T09:43:12,672 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-04T09:43:12,673 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:12,674 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:57868, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-04T09:43:12,675 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@70fbe73e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-04T09:43:12,676 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-04T09:43:12,677 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=84486a41f81c,32869,1733305391037, seqNum=-1] 2024-12-04T09:43:12,678 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-04T09:43:12,679 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58800, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-04T09:43:12,681 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=84486a41f81c,33655,1733305390799 2024-12-04T09:43:12,682 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-04T09:43:12,683 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is 84486a41f81c,33655,1733305390799 2024-12-04T09:43:12,683 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@55724972 2024-12-04T09:43:12,683 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-04T09:43:12,686 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:57870, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-04T09:43:12,687 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-04T09:43:12,689 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-04T09:43:12,691 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-04T09:43:12,691 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:12,691 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-04T09:43:12,693 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-04T09:43:12,693 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-04T09:43:12,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741837_1013 (size=392) 2024-12-04T09:43:12,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741837_1013 (size=392) 2024-12-04T09:43:12,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741837_1013 (size=392) 2024-12-04T09:43:12,707 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 49d7152fab23b4c2ec4a74c939e59a6c, NAME => 'TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0 2024-12-04T09:43:12,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741838_1014 (size=51) 2024-12-04T09:43:12,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741838_1014 (size=51) 2024-12-04T09:43:12,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741838_1014 (size=51) 2024-12-04T09:43:12,729 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:12,729 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 49d7152fab23b4c2ec4a74c939e59a6c, disabling compactions & flushes 2024-12-04T09:43:12,729 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:12,729 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:12,729 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. after waiting 0 ms 2024-12-04T09:43:12,730 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:12,730 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:12,730 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 49d7152fab23b4c2ec4a74c939e59a6c: Waiting for close lock at 1733305392729Disabling compacts and flushes for region at 1733305392729Disabling writes for close at 1733305392730 (+1 ms)Writing region close event to WAL at 1733305392730Closed at 1733305392730 2024-12-04T09:43:12,732 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-04T09:43:12,732 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733305392732"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733305392732"}]},"ts":"1733305392732"} 2024-12-04T09:43:12,735 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-04T09:43:12,737 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-04T09:43:12,737 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733305392737"}]},"ts":"1733305392737"} 2024-12-04T09:43:12,740 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-04T09:43:12,741 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {84486a41f81c=0} racks are {/default-rack=0} 2024-12-04T09:43:12,741 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-04T09:43:12,741 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-04T09:43:12,742 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-04T09:43:12,742 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-04T09:43:12,742 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-04T09:43:12,742 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-04T09:43:12,742 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-04T09:43:12,742 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-04T09:43:12,742 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-04T09:43:12,742 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-04T09:43:12,742 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=49d7152fab23b4c2ec4a74c939e59a6c, ASSIGN}] 2024-12-04T09:43:12,745 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=49d7152fab23b4c2ec4a74c939e59a6c, ASSIGN 2024-12-04T09:43:12,746 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=49d7152fab23b4c2ec4a74c939e59a6c, ASSIGN; state=OFFLINE, location=84486a41f81c,40111,1733305390950; forceNewPlan=false, retain=false 2024-12-04T09:43:12,797 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-04T09:43:12,897 INFO [84486a41f81c:33655 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-04T09:43:12,897 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=49d7152fab23b4c2ec4a74c939e59a6c, regionState=OPENING, regionLocation=84486a41f81c,40111,1733305390950 2024-12-04T09:43:12,901 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=49d7152fab23b4c2ec4a74c939e59a6c, ASSIGN because future has completed 2024-12-04T09:43:12,902 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 49d7152fab23b4c2ec4a74c939e59a6c, server=84486a41f81c,40111,1733305390950}] 2024-12-04T09:43:13,007 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-04T09:43:13,056 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-04T09:43:13,058 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35709, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-04T09:43:13,062 INFO [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,063 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 49d7152fab23b4c2ec4a74c939e59a6c, NAME => 'TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c.', STARTKEY => '', ENDKEY => ''} 2024-12-04T09:43:13,063 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,063 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-04T09:43:13,063 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,063 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,065 INFO [StoreOpener-49d7152fab23b4c2ec4a74c939e59a6c-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,066 INFO [StoreOpener-49d7152fab23b4c2ec4a74c939e59a6c-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 49d7152fab23b4c2ec4a74c939e59a6c columnFamilyName cf 2024-12-04T09:43:13,066 DEBUG [StoreOpener-49d7152fab23b4c2ec4a74c939e59a6c-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-04T09:43:13,067 INFO [StoreOpener-49d7152fab23b4c2ec4a74c939e59a6c-1 {}] regionserver.HStore(327): Store=49d7152fab23b4c2ec4a74c939e59a6c/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-04T09:43:13,067 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,068 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,069 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,069 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,069 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,072 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,075 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-04T09:43:13,076 INFO [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 49d7152fab23b4c2ec4a74c939e59a6c; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66087420, jitterRate=-0.015220701694488525}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-04T09:43:13,076 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,077 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 49d7152fab23b4c2ec4a74c939e59a6c: Running coprocessor pre-open hook at 1733305393063Writing region info on filesystem at 1733305393063Initializing all the Stores at 1733305393064 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733305393064Cleaning up temporary data from old regions at 1733305393070 (+6 ms)Running coprocessor post-open hooks at 1733305393076 (+6 ms)Region opened successfully at 1733305393077 (+1 ms) 2024-12-04T09:43:13,079 INFO [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c., pid=6, masterSystemTime=1733305393056 2024-12-04T09:43:13,082 DEBUG [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,082 INFO [RS_OPEN_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,083 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=49d7152fab23b4c2ec4a74c939e59a6c, regionState=OPEN, openSeqNum=2, regionLocation=84486a41f81c,40111,1733305390950 2024-12-04T09:43:13,086 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 49d7152fab23b4c2ec4a74c939e59a6c, server=84486a41f81c,40111,1733305390950 because future has completed 2024-12-04T09:43:13,092 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-04T09:43:13,092 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 49d7152fab23b4c2ec4a74c939e59a6c, server=84486a41f81c,40111,1733305390950 in 187 msec 2024-12-04T09:43:13,096 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-04T09:43:13,096 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=49d7152fab23b4c2ec4a74c939e59a6c, ASSIGN in 350 msec 2024-12-04T09:43:13,097 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-04T09:43:13,098 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733305393098"}]},"ts":"1733305393098"} 2024-12-04T09:43:13,101 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-04T09:43:13,103 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-04T09:43:13,106 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 416 msec 2024-12-04T09:43:13,317 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-04T09:43:13,318 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-04T09:43:13,318 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-04T09:43:13,318 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-04T09:43:13,321 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-04T09:43:13,321 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-04T09:43:13,321 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-04T09:43:13,325 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c., hostname=84486a41f81c,40111,1733305390950, seqNum=2] 2024-12-04T09:43:13,325 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-04T09:43:13,328 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56184, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-04T09:43:13,332 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-04T09:43:13,334 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-04T09:43:13,335 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-04T09:43:13,336 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-04T09:43:13,338 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-04T09:43:13,338 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-04T09:43:13,447 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-04T09:43:13,493 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=40111 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-04T09:43:13,494 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,494 INFO [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 49d7152fab23b4c2ec4a74c939e59a6c 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-04T09:43:13,516 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c/.tmp/cf/7da1eff9926442fb8c23ba473af9c003 is 36, key is row/cf:cq/1733305393329/Put/seqid=0 2024-12-04T09:43:13,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741839_1015 (size=4787) 2024-12-04T09:43:13,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741839_1015 (size=4787) 2024-12-04T09:43:13,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741839_1015 (size=4787) 2024-12-04T09:43:13,531 INFO [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c/.tmp/cf/7da1eff9926442fb8c23ba473af9c003 2024-12-04T09:43:13,540 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c/.tmp/cf/7da1eff9926442fb8c23ba473af9c003 as hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c/cf/7da1eff9926442fb8c23ba473af9c003 2024-12-04T09:43:13,552 INFO [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c/cf/7da1eff9926442fb8c23ba473af9c003, entries=1, sequenceid=5, filesize=4.7 K 2024-12-04T09:43:13,554 INFO [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 49d7152fab23b4c2ec4a74c939e59a6c in 60ms, sequenceid=5, compaction requested=false 2024-12-04T09:43:13,554 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 49d7152fab23b4c2ec4a74c939e59a6c: 2024-12-04T09:43:13,554 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,555 DEBUG [RS_FLUSH_OPERATIONS-regionserver/84486a41f81c:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-04T09:43:13,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-04T09:43:13,561 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-04T09:43:13,561 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 219 msec 2024-12-04T09:43:13,564 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 230 msec 2024-12-04T09:43:13,657 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33655 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-04T09:43:13,658 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-04T09:43:13,663 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-04T09:43:13,663 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-04T09:43:13,663 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:13,663 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:13,664 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:13,664 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-04T09:43:13,664 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-04T09:43:13,664 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1194308569, stopped=false 2024-12-04T09:43:13,664 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=84486a41f81c,33655,1733305390799 2024-12-04T09:43:13,703 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:13,703 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:13,703 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:13,703 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-04T09:43:13,703 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:13,703 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:13,703 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:13,703 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:13,704 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-04T09:43:13,704 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-04T09:43:13,705 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:13,705 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:13,705 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:13,705 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:13,705 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:13,705 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-04T09:43:13,705 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '84486a41f81c,40111,1733305390950' ***** 2024-12-04T09:43:13,705 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-04T09:43:13,705 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '84486a41f81c,39675,1733305390992' ***** 2024-12-04T09:43:13,705 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-04T09:43:13,705 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-04T09:43:13,705 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '84486a41f81c,32869,1733305391037' ***** 2024-12-04T09:43:13,705 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-04T09:43:13,706 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-04T09:43:13,706 INFO [RS:0;84486a41f81c:40111 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-04T09:43:13,706 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-04T09:43:13,706 INFO [RS:0;84486a41f81c:40111 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-04T09:43:13,706 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(3091): Received CLOSE for 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,706 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-04T09:43:13,706 INFO [RS:1;84486a41f81c:39675 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-04T09:43:13,706 INFO [RS:1;84486a41f81c:39675 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-04T09:43:13,706 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(959): stopping server 84486a41f81c,39675,1733305390992 2024-12-04T09:43:13,706 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(959): stopping server 84486a41f81c,40111,1733305390950 2024-12-04T09:43:13,706 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-04T09:43:13,707 INFO [RS:1;84486a41f81c:39675 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-04T09:43:13,707 INFO [RS:0;84486a41f81c:40111 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-04T09:43:13,707 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-04T09:43:13,707 INFO [RS:0;84486a41f81c:40111 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;84486a41f81c:40111. 2024-12-04T09:43:13,707 INFO [RS:1;84486a41f81c:39675 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;84486a41f81c:39675. 2024-12-04T09:43:13,707 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 49d7152fab23b4c2ec4a74c939e59a6c, disabling compactions & flushes 2024-12-04T09:43:13,707 INFO [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,707 DEBUG [RS:0;84486a41f81c:40111 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:13,707 DEBUG [RS:1;84486a41f81c:39675 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:13,707 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,707 DEBUG [RS:0;84486a41f81c:40111 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:13,707 DEBUG [RS:1;84486a41f81c:39675 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:13,707 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. after waiting 0 ms 2024-12-04T09:43:13,707 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(976): stopping server 84486a41f81c,39675,1733305390992; all regions closed. 2024-12-04T09:43:13,707 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-04T09:43:13,707 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,707 DEBUG [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1325): Online Regions={49d7152fab23b4c2ec4a74c939e59a6c=TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c.} 2024-12-04T09:43:13,708 DEBUG [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1351): Waiting on 49d7152fab23b4c2ec4a74c939e59a6c 2024-12-04T09:43:13,707 INFO [RS:2;84486a41f81c:32869 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-04T09:43:13,708 INFO [RS:2;84486a41f81c:32869 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-04T09:43:13,708 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,708 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(959): stopping server 84486a41f81c,32869,1733305391037 2024-12-04T09:43:13,708 INFO [RS:2;84486a41f81c:32869 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-04T09:43:13,708 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,708 INFO [RS:2;84486a41f81c:32869 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;84486a41f81c:32869. 2024-12-04T09:43:13,708 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,708 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,708 DEBUG [RS:2;84486a41f81c:32869 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-04T09:43:13,709 DEBUG [RS:2;84486a41f81c:32869 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:13,709 INFO [RS:2;84486a41f81c:32869 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-04T09:43:13,709 INFO [RS:2;84486a41f81c:32869 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-04T09:43:13,709 INFO [RS:2;84486a41f81c:32869 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-04T09:43:13,709 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-04T09:43:13,709 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-04T09:43:13,709 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,710 DEBUG [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-04T09:43:13,710 DEBUG [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-04T09:43:13,710 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-04T09:43:13,710 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-04T09:43:13,710 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-04T09:43:13,710 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-04T09:43:13,710 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-04T09:43:13,710 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-04T09:43:13,714 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/default/TestHBaseWalOnEC/49d7152fab23b4c2ec4a74c939e59a6c/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-04T09:43:13,715 INFO [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,716 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 49d7152fab23b4c2ec4a74c939e59a6c: Waiting for close lock at 1733305393707Running coprocessor pre-close hooks at 1733305393707Disabling compacts and flushes for region at 1733305393707Disabling writes for close at 1733305393707Writing region close event to WAL at 1733305393708 (+1 ms)Running coprocessor post-close hooks at 1733305393715 (+7 ms)Closed at 1733305393715 2024-12-04T09:43:13,716 DEBUG [RS_CLOSE_REGION-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c. 2024-12-04T09:43:13,719 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741834_1010 (size=93) 2024-12-04T09:43:13,719 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741834_1010 (size=93) 2024-12-04T09:43:13,719 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741834_1010 (size=93) 2024-12-04T09:43:13,730 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/info/564c7e0d163940a18ca2a8453fee2e8b is 153, key is TestHBaseWalOnEC,,1733305392686.49d7152fab23b4c2ec4a74c939e59a6c./info:regioninfo/1733305393083/Put/seqid=0 2024-12-04T09:43:13,731 INFO [regionserver/84486a41f81c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:13,732 INFO [regionserver/84486a41f81c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:13,732 INFO [regionserver/84486a41f81c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:13,737 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741840_1016 (size=6637) 2024-12-04T09:43:13,737 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741840_1016 (size=6637) 2024-12-04T09:43:13,738 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741840_1016 (size=6637) 2024-12-04T09:43:13,739 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/info/564c7e0d163940a18ca2a8453fee2e8b 2024-12-04T09:43:13,762 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/ns/263378b81f054eb5a67f0a54365107bc is 43, key is default/ns:d/1733305392551/Put/seqid=0 2024-12-04T09:43:13,772 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741841_1017 (size=5153) 2024-12-04T09:43:13,773 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741841_1017 (size=5153) 2024-12-04T09:43:13,774 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741841_1017 (size=5153) 2024-12-04T09:43:13,774 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/ns/263378b81f054eb5a67f0a54365107bc 2024-12-04T09:43:13,803 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/table/c6f6e4ff759c44adab919471781ddbc4 is 52, key is TestHBaseWalOnEC/table:state/1733305393098/Put/seqid=0 2024-12-04T09:43:13,810 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741842_1018 (size=5249) 2024-12-04T09:43:13,810 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741842_1018 (size=5249) 2024-12-04T09:43:13,811 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741842_1018 (size=5249) 2024-12-04T09:43:13,811 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/table/c6f6e4ff759c44adab919471781ddbc4 2024-12-04T09:43:13,819 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/info/564c7e0d163940a18ca2a8453fee2e8b as hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/info/564c7e0d163940a18ca2a8453fee2e8b 2024-12-04T09:43:13,827 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/info/564c7e0d163940a18ca2a8453fee2e8b, entries=10, sequenceid=11, filesize=6.5 K 2024-12-04T09:43:13,828 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/ns/263378b81f054eb5a67f0a54365107bc as hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/ns/263378b81f054eb5a67f0a54365107bc 2024-12-04T09:43:13,836 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/ns/263378b81f054eb5a67f0a54365107bc, entries=2, sequenceid=11, filesize=5.0 K 2024-12-04T09:43:13,837 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/.tmp/table/c6f6e4ff759c44adab919471781ddbc4 as hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/table/c6f6e4ff759c44adab919471781ddbc4 2024-12-04T09:43:13,845 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/table/c6f6e4ff759c44adab919471781ddbc4, entries=2, sequenceid=11, filesize=5.1 K 2024-12-04T09:43:13,846 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 136ms, sequenceid=11, compaction requested=false 2024-12-04T09:43:13,851 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-04T09:43:13,852 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-04T09:43:13,852 INFO [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-04T09:43:13,852 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733305393710Running coprocessor pre-close hooks at 1733305393710Disabling compacts and flushes for region at 1733305393710Disabling writes for close at 1733305393710Obtaining lock to block concurrent updates at 1733305393710Preparing flush snapshotting stores in 1588230740 at 1733305393710Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733305393711 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733305393712 (+1 ms)Flushing 1588230740/info: creating writer at 1733305393712Flushing 1588230740/info: appending metadata at 1733305393729 (+17 ms)Flushing 1588230740/info: closing flushed file at 1733305393729Flushing 1588230740/ns: creating writer at 1733305393748 (+19 ms)Flushing 1588230740/ns: appending metadata at 1733305393762 (+14 ms)Flushing 1588230740/ns: closing flushed file at 1733305393762Flushing 1588230740/table: creating writer at 1733305393784 (+22 ms)Flushing 1588230740/table: appending metadata at 1733305393803 (+19 ms)Flushing 1588230740/table: closing flushed file at 1733305393803Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3366a653: reopening flushed file at 1733305393818 (+15 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2e2332f2: reopening flushed file at 1733305393827 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@24993f38: reopening flushed file at 1733305393836 (+9 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 136ms, sequenceid=11, compaction requested=false at 1733305393846 (+10 ms)Writing region close event to WAL at 1733305393847 (+1 ms)Running coprocessor post-close hooks at 1733305393852 (+5 ms)Closed at 1733305393852 2024-12-04T09:43:13,852 DEBUG [RS_CLOSE_META-regionserver/84486a41f81c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-04T09:43:13,908 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(976): stopping server 84486a41f81c,40111,1733305390950; all regions closed. 2024-12-04T09:43:13,908 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,908 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,909 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,909 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,909 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,910 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(976): stopping server 84486a41f81c,32869,1733305391037; all regions closed. 2024-12-04T09:43:13,910 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,910 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,910 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,911 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,911 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741833_1009 (size=1298) 2024-12-04T09:43:13,912 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741833_1009 (size=1298) 2024-12-04T09:43:13,913 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741833_1009 (size=1298) 2024-12-04T09:43:13,916 DEBUG [RS:0;84486a41f81c:40111 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/oldWALs 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 84486a41f81c%2C40111%2C1733305390950:(num 1733305391766) 2024-12-04T09:43:13,916 DEBUG [RS:0;84486a41f81c:40111 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] hbase.ChoreService(370): Chore service for: regionserver/84486a41f81c:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-04T09:43:13,916 INFO [regionserver/84486a41f81c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-04T09:43:13,916 INFO [RS:0;84486a41f81c:40111 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:40111 2024-12-04T09:43:13,923 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/84486a41f81c,40111,1733305390950 2024-12-04T09:43:13,923 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:13,923 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [84486a41f81c,40111,1733305390950] 2024-12-04T09:43:13,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741836_1012 (size=2751) 2024-12-04T09:43:13,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741836_1012 (size=2751) 2024-12-04T09:43:13,925 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741836_1012 (size=2751) 2024-12-04T09:43:13,928 DEBUG [RS:2;84486a41f81c:32869 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/oldWALs 2024-12-04T09:43:13,928 INFO [RS:2;84486a41f81c:32869 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 84486a41f81c%2C32869%2C1733305391037.meta:.meta(num 1733305392478) 2024-12-04T09:43:13,928 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,928 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,928 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,928 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,929 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:13,931 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/84486a41f81c,40111,1733305390950 already deleted, retry=false 2024-12-04T09:43:13,931 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 84486a41f81c,40111,1733305390950 expired; onlineServers=2 2024-12-04T09:43:13,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741835_1011 (size=93) 2024-12-04T09:43:13,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741835_1011 (size=93) 2024-12-04T09:43:13,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741835_1011 (size=93) 2024-12-04T09:43:13,934 DEBUG [RS:2;84486a41f81c:32869 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/oldWALs 2024-12-04T09:43:13,934 INFO [RS:2;84486a41f81c:32869 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 84486a41f81c%2C32869%2C1733305391037:(num 1733305391773) 2024-12-04T09:43:13,934 DEBUG [RS:2;84486a41f81c:32869 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:13,934 INFO [RS:2;84486a41f81c:32869 {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:13,934 INFO [RS:2;84486a41f81c:32869 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-04T09:43:13,934 INFO [RS:2;84486a41f81c:32869 {}] hbase.ChoreService(370): Chore service for: regionserver/84486a41f81c:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-04T09:43:13,934 INFO [RS:2;84486a41f81c:32869 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-04T09:43:13,934 INFO [regionserver/84486a41f81c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-04T09:43:13,934 INFO [RS:2;84486a41f81c:32869 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:32869 2024-12-04T09:43:13,934 INFO [RS:0;84486a41f81c:40111 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-04T09:43:13,953 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:13,953 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/84486a41f81c,32869,1733305391037 2024-12-04T09:43:13,953 INFO [RS:2;84486a41f81c:32869 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-04T09:43:13,962 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [84486a41f81c,32869,1733305391037] 2024-12-04T09:43:13,969 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/84486a41f81c,32869,1733305391037 already deleted, retry=false 2024-12-04T09:43:13,970 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 84486a41f81c,32869,1733305391037 expired; onlineServers=1 2024-12-04T09:43:14,045 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:14,045 INFO [RS:0;84486a41f81c:40111 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-04T09:43:14,045 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40111-0x101a102f92f0001, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:14,045 INFO [RS:0;84486a41f81c:40111 {}] regionserver.HRegionServer(1031): Exiting; stopping=84486a41f81c,40111,1733305390950; zookeeper connection closed. 2024-12-04T09:43:14,046 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@268b907b {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@268b907b 2024-12-04T09:43:14,062 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:14,062 INFO [RS:2;84486a41f81c:32869 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-04T09:43:14,062 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32869-0x101a102f92f0003, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:14,062 INFO [RS:2;84486a41f81c:32869 {}] regionserver.HRegionServer(1031): Exiting; stopping=84486a41f81c,32869,1733305391037; zookeeper connection closed. 2024-12-04T09:43:14,062 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@42ddda55 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@42ddda55 2024-12-04T09:43:14,121 DEBUG [RS:1;84486a41f81c:39675 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/oldWALs 2024-12-04T09:43:14,121 INFO [RS:1;84486a41f81c:39675 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 84486a41f81c%2C39675%2C1733305390992:(num 1733305391770) 2024-12-04T09:43:14,121 DEBUG [RS:1;84486a41f81c:39675 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-04T09:43:14,121 INFO [RS:1;84486a41f81c:39675 {}] regionserver.LeaseManager(133): Closed leases 2024-12-04T09:43:14,121 INFO [RS:1;84486a41f81c:39675 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-04T09:43:14,121 INFO [RS:1;84486a41f81c:39675 {}] hbase.ChoreService(370): Chore service for: regionserver/84486a41f81c:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-04T09:43:14,122 INFO [RS:1;84486a41f81c:39675 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-04T09:43:14,122 INFO [RS:1;84486a41f81c:39675 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-04T09:43:14,122 INFO [regionserver/84486a41f81c:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-04T09:43:14,122 INFO [RS:1;84486a41f81c:39675 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-04T09:43:14,122 INFO [RS:1;84486a41f81c:39675 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-04T09:43:14,122 INFO [RS:1;84486a41f81c:39675 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39675 2024-12-04T09:43:14,131 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/84486a41f81c,39675,1733305390992 2024-12-04T09:43:14,131 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-04T09:43:14,131 INFO [RS:1;84486a41f81c:39675 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-04T09:43:14,138 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-04T09:43:14,139 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-04T09:43:14,139 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [84486a41f81c,39675,1733305390992] 2024-12-04T09:43:14,148 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/84486a41f81c,39675,1733305390992 already deleted, retry=false 2024-12-04T09:43:14,148 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 84486a41f81c,39675,1733305390992 expired; onlineServers=0 2024-12-04T09:43:14,148 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '84486a41f81c,33655,1733305390799' ***** 2024-12-04T09:43:14,148 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-04T09:43:14,148 INFO [M:0;84486a41f81c:33655 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-04T09:43:14,148 INFO [M:0;84486a41f81c:33655 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-04T09:43:14,148 DEBUG [M:0;84486a41f81c:33655 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-04T09:43:14,148 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-04T09:43:14,148 DEBUG [M:0;84486a41f81c:33655 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-04T09:43:14,148 DEBUG [master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.small.0-1733305391451 {}] cleaner.HFileCleaner(306): Exit Thread[master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.small.0-1733305391451,5,FailOnTimeoutGroup] 2024-12-04T09:43:14,148 DEBUG [master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.large.0-1733305391451 {}] cleaner.HFileCleaner(306): Exit Thread[master/84486a41f81c:0:becomeActiveMaster-HFileCleaner.large.0-1733305391451,5,FailOnTimeoutGroup] 2024-12-04T09:43:14,148 INFO [M:0;84486a41f81c:33655 {}] hbase.ChoreService(370): Chore service for: master/84486a41f81c:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-04T09:43:14,148 INFO [M:0;84486a41f81c:33655 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-04T09:43:14,149 DEBUG [M:0;84486a41f81c:33655 {}] master.HMaster(1795): Stopping service threads 2024-12-04T09:43:14,149 INFO [M:0;84486a41f81c:33655 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-04T09:43:14,149 INFO [M:0;84486a41f81c:33655 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-04T09:43:14,149 INFO [M:0;84486a41f81c:33655 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-04T09:43:14,149 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-04T09:43:14,156 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-04T09:43:14,156 DEBUG [M:0;84486a41f81c:33655 {}] zookeeper.ZKUtil(347): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-04T09:43:14,156 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-04T09:43:14,156 WARN [M:0;84486a41f81c:33655 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-04T09:43:14,157 INFO [M:0;84486a41f81c:33655 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/.lastflushedseqids 2024-12-04T09:43:14,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741843_1019 (size=127) 2024-12-04T09:43:14,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741843_1019 (size=127) 2024-12-04T09:43:14,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741843_1019 (size=127) 2024-12-04T09:43:14,174 INFO [M:0;84486a41f81c:33655 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-04T09:43:14,174 INFO [M:0;84486a41f81c:33655 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-04T09:43:14,174 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-04T09:43:14,174 INFO [M:0;84486a41f81c:33655 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:14,174 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:14,174 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-04T09:43:14,174 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:14,175 INFO [M:0;84486a41f81c:33655 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.82 KB heapSize=34.11 KB 2024-12-04T09:43:14,192 DEBUG [M:0;84486a41f81c:33655 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f945483cf2084b5f99ac539426a4416e is 82, key is hbase:meta,,1/info:regioninfo/1733305392521/Put/seqid=0 2024-12-04T09:43:14,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741844_1020 (size=5672) 2024-12-04T09:43:14,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741844_1020 (size=5672) 2024-12-04T09:43:14,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741844_1020 (size=5672) 2024-12-04T09:43:14,205 INFO [M:0;84486a41f81c:33655 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f945483cf2084b5f99ac539426a4416e 2024-12-04T09:43:14,226 DEBUG [M:0;84486a41f81c:33655 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/303b216e626e44b2a88ad029f24dc787 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733305393105/Put/seqid=0 2024-12-04T09:43:14,234 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741845_1021 (size=6438) 2024-12-04T09:43:14,234 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741845_1021 (size=6438) 2024-12-04T09:43:14,234 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741845_1021 (size=6438) 2024-12-04T09:43:14,235 INFO [M:0;84486a41f81c:33655 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/303b216e626e44b2a88ad029f24dc787 2024-12-04T09:43:14,240 INFO [RS:1;84486a41f81c:39675 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-04T09:43:14,240 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:14,240 INFO [RS:1;84486a41f81c:39675 {}] regionserver.HRegionServer(1031): Exiting; stopping=84486a41f81c,39675,1733305390992; zookeeper connection closed. 2024-12-04T09:43:14,240 DEBUG [pool-330-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39675-0x101a102f92f0002, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:14,240 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7975de25 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7975de25 2024-12-04T09:43:14,240 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-04T09:43:14,260 DEBUG [M:0;84486a41f81c:33655 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/82e8e22616954f1e884c185bafa8d84a is 69, key is 84486a41f81c,32869,1733305391037/rs:state/1733305391562/Put/seqid=0 2024-12-04T09:43:14,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741846_1022 (size=5294) 2024-12-04T09:43:14,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741846_1022 (size=5294) 2024-12-04T09:43:14,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741846_1022 (size=5294) 2024-12-04T09:43:14,269 INFO [M:0;84486a41f81c:33655 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/82e8e22616954f1e884c185bafa8d84a 2024-12-04T09:43:14,277 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f945483cf2084b5f99ac539426a4416e as hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/f945483cf2084b5f99ac539426a4416e 2024-12-04T09:43:14,285 INFO [M:0;84486a41f81c:33655 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/f945483cf2084b5f99ac539426a4416e, entries=8, sequenceid=72, filesize=5.5 K 2024-12-04T09:43:14,286 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/303b216e626e44b2a88ad029f24dc787 as hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/303b216e626e44b2a88ad029f24dc787 2024-12-04T09:43:14,293 INFO [M:0;84486a41f81c:33655 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/303b216e626e44b2a88ad029f24dc787, entries=8, sequenceid=72, filesize=6.3 K 2024-12-04T09:43:14,294 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/82e8e22616954f1e884c185bafa8d84a as hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/82e8e22616954f1e884c185bafa8d84a 2024-12-04T09:43:14,300 INFO [M:0;84486a41f81c:33655 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:34577/user/jenkins/test-data/a1da783e-a634-1c61-9bf2-3c0df944d9b0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/82e8e22616954f1e884c185bafa8d84a, entries=3, sequenceid=72, filesize=5.2 K 2024-12-04T09:43:14,302 INFO [M:0;84486a41f81c:33655 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.82 KB/27462, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 128ms, sequenceid=72, compaction requested=false 2024-12-04T09:43:14,307 INFO [M:0;84486a41f81c:33655 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-04T09:43:14,307 DEBUG [M:0;84486a41f81c:33655 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733305394174Disabling compacts and flushes for region at 1733305394174Disabling writes for close at 1733305394174Obtaining lock to block concurrent updates at 1733305394175 (+1 ms)Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733305394175Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27462, getHeapSize=34864, getOffHeapSize=0, getCellsCount=85 at 1733305394175Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733305394176 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733305394176Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733305394192 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733305394192Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733305394211 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733305394225 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733305394226 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733305394242 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733305394258 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733305394258Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@56d81b1a: reopening flushed file at 1733305394275 (+17 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6286fbc4: reopening flushed file at 1733305394285 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@35b19489: reopening flushed file at 1733305394293 (+8 ms)Finished flush of dataSize ~26.82 KB/27462, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 128ms, sequenceid=72, compaction requested=false at 1733305394302 (+9 ms)Writing region close event to WAL at 1733305394307 (+5 ms)Closed at 1733305394307 2024-12-04T09:43:14,307 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:14,307 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:14,307 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:14,308 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:14,308 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-04T09:43:14,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41519 is added to blk_1073741830_1006 (size=32665) 2024-12-04T09:43:14,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45145 is added to blk_1073741830_1006 (size=32665) 2024-12-04T09:43:14,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43879 is added to blk_1073741830_1006 (size=32665) 2024-12-04T09:43:14,312 INFO [M:0;84486a41f81c:33655 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-04T09:43:14,312 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-04T09:43:14,312 INFO [M:0;84486a41f81c:33655 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:33655 2024-12-04T09:43:14,312 INFO [M:0;84486a41f81c:33655 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-04T09:43:14,423 INFO [M:0;84486a41f81c:33655 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-04T09:43:14,423 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:14,423 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33655-0x101a102f92f0000, quorum=127.0.0.1:51778, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-04T09:43:14,427 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@44672b71{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:14,428 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@36a9ca95{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-04T09:43:14,428 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-04T09:43:14,428 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@438bc7ce{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-04T09:43:14,428 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4c77de1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir/,STOPPED} 2024-12-04T09:43:14,429 WARN [BP-1682927199-172.17.0.2-1733305388802 heartbeating to localhost/127.0.0.1:34577 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-04T09:43:14,429 WARN [BP-1682927199-172.17.0.2-1733305388802 heartbeating to localhost/127.0.0.1:34577 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1682927199-172.17.0.2-1733305388802 (Datanode Uuid efc6e4c9-c5aa-4888-9f32-4a1d33a8ec16) service to localhost/127.0.0.1:34577 2024-12-04T09:43:14,430 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data5/current/BP-1682927199-172.17.0.2-1733305388802 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:14,430 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data6/current/BP-1682927199-172.17.0.2-1733305388802 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:14,430 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-04T09:43:14,430 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-04T09:43:14,430 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-04T09:43:14,438 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@30bdc6f8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:14,443 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5547eae9{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-04T09:43:14,443 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-04T09:43:14,443 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1b0441b5{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-04T09:43:14,444 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4e5afbc4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir/,STOPPED} 2024-12-04T09:43:14,445 WARN [BP-1682927199-172.17.0.2-1733305388802 heartbeating to localhost/127.0.0.1:34577 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-04T09:43:14,445 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-04T09:43:14,445 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-04T09:43:14,445 WARN [BP-1682927199-172.17.0.2-1733305388802 heartbeating to localhost/127.0.0.1:34577 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1682927199-172.17.0.2-1733305388802 (Datanode Uuid cb5ce412-3855-4d0f-98b2-54019b9cffa1) service to localhost/127.0.0.1:34577 2024-12-04T09:43:14,446 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data3/current/BP-1682927199-172.17.0.2-1733305388802 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:14,446 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data4/current/BP-1682927199-172.17.0.2-1733305388802 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:14,446 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-04T09:43:14,449 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7eeef71e{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-04T09:43:14,449 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@70feba44{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-04T09:43:14,449 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-04T09:43:14,449 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@20a0e688{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-04T09:43:14,449 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@61d23bc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir/,STOPPED} 2024-12-04T09:43:14,450 WARN [BP-1682927199-172.17.0.2-1733305388802 heartbeating to localhost/127.0.0.1:34577 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-04T09:43:14,450 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-04T09:43:14,450 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-04T09:43:14,450 WARN [BP-1682927199-172.17.0.2-1733305388802 heartbeating to localhost/127.0.0.1:34577 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1682927199-172.17.0.2-1733305388802 (Datanode Uuid ea2e42a4-93fc-49c3-ab65-9ee31c501e86) service to localhost/127.0.0.1:34577 2024-12-04T09:43:14,451 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data1/current/BP-1682927199-172.17.0.2-1733305388802 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:14,451 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/cluster_7e904ddf-3935-91ea-2ce0-9d36e24a04b8/data/data2/current/BP-1682927199-172.17.0.2-1733305388802 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-04T09:43:14,451 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-04T09:43:14,457 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3c97821d{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-04T09:43:14,458 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@544c0dd2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-04T09:43:14,458 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-04T09:43:14,458 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@62802e4c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-04T09:43:14,458 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@16eaa68d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/ee979d88-df78-6bf0-6572-d18bfefbf864/hadoop.log.dir/,STOPPED} 2024-12-04T09:43:14,465 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-04T09:43:14,488 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-04T09:43:14,495 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=152 (was 92) - Thread LEAK? -, OpenFileDescriptor=518 (was 437) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=347 (was 316) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=10218 (was 10400)