2024-11-11 03:15:23,076 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-11-11 03:15:23,088 main DEBUG Took 0.010111 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-11-11 03:15:23,088 main DEBUG PluginManager 'Core' found 129 plugins 2024-11-11 03:15:23,088 main DEBUG PluginManager 'Level' found 0 plugins 2024-11-11 03:15:23,089 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-11-11 03:15:23,090 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,100 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-11-11 03:15:23,117 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,118 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,119 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,119 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,120 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,120 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,121 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,121 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,121 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,122 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,122 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,123 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,123 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,123 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,124 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,124 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,125 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,125 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,125 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,125 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,126 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,126 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,126 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,127 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-11 03:15:23,127 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,127 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-11-11 03:15:23,129 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-11 03:15:23,130 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-11-11 03:15:23,131 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-11-11 03:15:23,132 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-11-11 03:15:23,133 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-11-11 03:15:23,133 main DEBUG PluginManager 'Converter' found 47 plugins 2024-11-11 03:15:23,141 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-11-11 03:15:23,144 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-11-11 03:15:23,145 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-11-11 03:15:23,146 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-11-11 03:15:23,146 main DEBUG createAppenders(={Console}) 2024-11-11 03:15:23,147 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-11-11 03:15:23,147 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-11-11 03:15:23,147 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-11-11 03:15:23,148 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-11-11 03:15:23,148 main DEBUG OutputStream closed 2024-11-11 03:15:23,148 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-11-11 03:15:23,148 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-11-11 03:15:23,149 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-11-11 03:15:23,225 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-11-11 03:15:23,228 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-11-11 03:15:23,229 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-11-11 03:15:23,230 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-11-11 03:15:23,231 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-11-11 03:15:23,232 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-11-11 03:15:23,232 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-11-11 03:15:23,233 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-11-11 03:15:23,233 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-11-11 03:15:23,234 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-11-11 03:15:23,234 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-11-11 03:15:23,235 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-11-11 03:15:23,235 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-11-11 03:15:23,235 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-11-11 03:15:23,236 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-11-11 03:15:23,236 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-11-11 03:15:23,237 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-11-11 03:15:23,238 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-11-11 03:15:23,241 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-11-11 03:15:23,241 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-logging/target/hbase-logging-3.0.0-beta-2-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-11-11 03:15:23,242 main DEBUG Shutdown hook enabled. Registering a new one. 2024-11-11 03:15:23,243 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-11-11T03:15:23,266 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-11-11 03:15:23,269 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-11-11 03:15:23,269 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-11-11T03:15:23,507 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e 2024-11-11T03:15:23,531 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4, deleteOnExit=true 2024-11-11T03:15:23,532 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/test.cache.data in system properties and HBase conf 2024-11-11T03:15:23,533 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.tmp.dir in system properties and HBase conf 2024-11-11T03:15:23,534 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir in system properties and HBase conf 2024-11-11T03:15:23,534 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/mapreduce.cluster.local.dir in system properties and HBase conf 2024-11-11T03:15:23,535 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-11-11T03:15:23,535 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-11-11T03:15:23,639 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-11-11T03:15:23,729 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-11-11T03:15:23,732 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-11-11T03:15:23,732 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-11-11T03:15:23,733 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-11-11T03:15:23,733 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-11T03:15:23,734 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-11-11T03:15:23,734 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-11-11T03:15:23,735 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-11T03:15:23,735 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-11T03:15:23,735 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-11-11T03:15:23,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/nfs.dump.dir in system properties and HBase conf 2024-11-11T03:15:23,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/java.io.tmpdir in system properties and HBase conf 2024-11-11T03:15:23,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-11T03:15:23,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-11-11T03:15:23,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-11-11T03:15:24,552 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-11-11T03:15:24,634 INFO [Time-limited test {}] log.Log(170): Logging initialized @2194ms to org.eclipse.jetty.util.log.Slf4jLog 2024-11-11T03:15:24,712 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:24,774 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-11T03:15:24,796 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-11T03:15:24,796 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-11T03:15:24,798 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-11T03:15:24,810 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:24,813 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir/,AVAILABLE} 2024-11-11T03:15:24,814 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-11T03:15:25,028 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@76e4c45c{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/java.io.tmpdir/jetty-localhost-40303-hadoop-hdfs-3_4_1-tests_jar-_-any-4949742369734033064/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-11T03:15:25,041 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:40303} 2024-11-11T03:15:25,041 INFO [Time-limited test {}] server.Server(415): Started @2601ms 2024-11-11T03:15:25,420 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:25,426 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-11T03:15:25,427 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-11T03:15:25,427 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-11T03:15:25,428 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-11T03:15:25,429 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@550154bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir/,AVAILABLE} 2024-11-11T03:15:25,429 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1a2478ad{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-11T03:15:25,549 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4839957b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/java.io.tmpdir/jetty-localhost-34313-hadoop-hdfs-3_4_1-tests_jar-_-any-3643836459967037711/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:25,550 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5306f615{HTTP/1.1, (http/1.1)}{localhost:34313} 2024-11-11T03:15:25,550 INFO [Time-limited test {}] server.Server(415): Started @3110ms 2024-11-11T03:15:25,606 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-11T03:15:25,722 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:25,727 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-11T03:15:25,728 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-11T03:15:25,728 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-11T03:15:25,728 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-11T03:15:25,730 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6463ad04{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir/,AVAILABLE} 2024-11-11T03:15:25,730 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7fa8fa5c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-11T03:15:25,850 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1c6b8f01{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/java.io.tmpdir/jetty-localhost-44685-hadoop-hdfs-3_4_1-tests_jar-_-any-4555160600557998205/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:25,851 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11f28dd2{HTTP/1.1, (http/1.1)}{localhost:44685} 2024-11-11T03:15:25,851 INFO [Time-limited test {}] server.Server(415): Started @3412ms 2024-11-11T03:15:25,854 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-11T03:15:25,894 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:25,898 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-11T03:15:25,899 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-11T03:15:25,900 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-11T03:15:25,900 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-11T03:15:25,901 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@c62369b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir/,AVAILABLE} 2024-11-11T03:15:25,901 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@24f92c39{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-11T03:15:26,028 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2e59159d{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/java.io.tmpdir/jetty-localhost-35093-hadoop-hdfs-3_4_1-tests_jar-_-any-4199619296735194053/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:26,029 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@a8e922f{HTTP/1.1, (http/1.1)}{localhost:35093} 2024-11-11T03:15:26,029 INFO [Time-limited test {}] server.Server(415): Started @3590ms 2024-11-11T03:15:26,032 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-11T03:15:26,061 WARN [Thread-111 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data4/current/BP-1584767127-172.17.0.2-1731294924293/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:26,061 WARN [Thread-112 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data2/current/BP-1584767127-172.17.0.2-1731294924293/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:26,061 WARN [Thread-109 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data3/current/BP-1584767127-172.17.0.2-1731294924293/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:26,061 WARN [Thread-110 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data1/current/BP-1584767127-172.17.0.2-1731294924293/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:26,115 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-11T03:15:26,116 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-11T03:15:26,175 WARN [Thread-139 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data5/current/BP-1584767127-172.17.0.2-1731294924293/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:26,176 WARN [Thread-140 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data6/current/BP-1584767127-172.17.0.2-1731294924293/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:26,192 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x1cb6035bfda8fbd6 with lease ID 0x61787b2f1deb0874: Processing first storage report for DS-28348765-dcc6-4606-a18e-329f4b7e00f5 from datanode DatanodeRegistration(127.0.0.1:45271, datanodeUuid=edf79a88-1426-48ec-8141-065e335b0491, infoPort=36633, infoSecurePort=0, ipcPort=38833, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293) 2024-11-11T03:15:26,193 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x1cb6035bfda8fbd6 with lease ID 0x61787b2f1deb0874: from storage DS-28348765-dcc6-4606-a18e-329f4b7e00f5 node DatanodeRegistration(127.0.0.1:45271, datanodeUuid=edf79a88-1426-48ec-8141-065e335b0491, infoPort=36633, infoSecurePort=0, ipcPort=38833, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2024-11-11T03:15:26,193 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xfbcc43e8d75e1f65 with lease ID 0x61787b2f1deb0875: Processing first storage report for DS-11f7bbb6-ae48-466e-9dc3-94891c1bd02e from datanode DatanodeRegistration(127.0.0.1:34821, datanodeUuid=0355b2ca-8a6f-4761-9367-123eaf18ca40, infoPort=45315, infoSecurePort=0, ipcPort=40535, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293) 2024-11-11T03:15:26,194 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xfbcc43e8d75e1f65 with lease ID 0x61787b2f1deb0875: from storage DS-11f7bbb6-ae48-466e-9dc3-94891c1bd02e node DatanodeRegistration(127.0.0.1:34821, datanodeUuid=0355b2ca-8a6f-4761-9367-123eaf18ca40, infoPort=45315, infoSecurePort=0, ipcPort=40535, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-11T03:15:26,194 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x1cb6035bfda8fbd6 with lease ID 0x61787b2f1deb0874: Processing first storage report for DS-b665792a-5bdb-43b3-a8be-8410dae3916f from datanode DatanodeRegistration(127.0.0.1:45271, datanodeUuid=edf79a88-1426-48ec-8141-065e335b0491, infoPort=36633, infoSecurePort=0, ipcPort=38833, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293) 2024-11-11T03:15:26,194 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x1cb6035bfda8fbd6 with lease ID 0x61787b2f1deb0874: from storage DS-b665792a-5bdb-43b3-a8be-8410dae3916f node DatanodeRegistration(127.0.0.1:45271, datanodeUuid=edf79a88-1426-48ec-8141-065e335b0491, infoPort=36633, infoSecurePort=0, ipcPort=38833, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-11T03:15:26,195 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xfbcc43e8d75e1f65 with lease ID 0x61787b2f1deb0875: Processing first storage report for DS-3413b80c-68eb-4906-846d-8f3a3cb32f99 from datanode DatanodeRegistration(127.0.0.1:34821, datanodeUuid=0355b2ca-8a6f-4761-9367-123eaf18ca40, infoPort=45315, infoSecurePort=0, ipcPort=40535, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293) 2024-11-11T03:15:26,195 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xfbcc43e8d75e1f65 with lease ID 0x61787b2f1deb0875: from storage DS-3413b80c-68eb-4906-846d-8f3a3cb32f99 node DatanodeRegistration(127.0.0.1:34821, datanodeUuid=0355b2ca-8a6f-4761-9367-123eaf18ca40, infoPort=45315, infoSecurePort=0, ipcPort=40535, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-11T03:15:26,210 WARN [Thread-115 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-11T03:15:26,215 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdca2294e1753f9e7 with lease ID 0x61787b2f1deb0876: Processing first storage report for DS-713ba5f4-b3a0-49fe-a9cd-2acddc4f9ee6 from datanode DatanodeRegistration(127.0.0.1:44249, datanodeUuid=99217cbf-e2d6-4c92-ae21-66ce2577d1c7, infoPort=37915, infoSecurePort=0, ipcPort=40635, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293) 2024-11-11T03:15:26,215 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdca2294e1753f9e7 with lease ID 0x61787b2f1deb0876: from storage DS-713ba5f4-b3a0-49fe-a9cd-2acddc4f9ee6 node DatanodeRegistration(127.0.0.1:44249, datanodeUuid=99217cbf-e2d6-4c92-ae21-66ce2577d1c7, infoPort=37915, infoSecurePort=0, ipcPort=40635, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-11T03:15:26,215 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdca2294e1753f9e7 with lease ID 0x61787b2f1deb0876: Processing first storage report for DS-55fc5b2e-b3fd-43f4-8815-8dd9e0c87477 from datanode DatanodeRegistration(127.0.0.1:44249, datanodeUuid=99217cbf-e2d6-4c92-ae21-66ce2577d1c7, infoPort=37915, infoSecurePort=0, ipcPort=40635, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293) 2024-11-11T03:15:26,216 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdca2294e1753f9e7 with lease ID 0x61787b2f1deb0876: from storage DS-55fc5b2e-b3fd-43f4-8815-8dd9e0c87477 node DatanodeRegistration(127.0.0.1:44249, datanodeUuid=99217cbf-e2d6-4c92-ae21-66ce2577d1c7, infoPort=37915, infoSecurePort=0, ipcPort=40635, storageInfo=lv=-57;cid=testClusterID;nsid=1005398987;c=1731294924293), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-11T03:15:26,461 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e 2024-11-11T03:15:26,536 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-11-11T03:15:26,591 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=159, OpenFileDescriptor=391, MaxFileDescriptor=1048576, SystemLoadAverage=17, ProcessCount=11, AvailableMemoryMB=8329 2024-11-11T03:15:26,593 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-11-11T03:15:26,600 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-11-11T03:15:26,674 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/zookeeper_0, clientPort=54485, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-11-11T03:15:26,684 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=54485 2024-11-11T03:15:26,711 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:26,715 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:26,809 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:26,809 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:26,856 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:53020 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:45271:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53020 dst: /127.0.0.1:45271 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:26,879 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775792_1002 (size=7) 2024-11-11T03:15:27,275 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:27,286 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939 with version=8 2024-11-11T03:15:27,286 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/hbase-staging 2024-11-11T03:15:27,384 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-11-11T03:15:27,632 INFO [Time-limited test {}] client.ConnectionUtils(128): master/59e6b60f6143:0 server-side Connection retries=45 2024-11-11T03:15:27,643 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:27,644 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:27,649 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-11T03:15:27,649 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:27,649 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-11T03:15:27,784 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-11-11T03:15:27,847 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-11-11T03:15:27,856 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-11-11T03:15:27,860 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-11T03:15:27,888 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 28972 (auto-detected) 2024-11-11T03:15:27,889 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-11-11T03:15:27,908 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:42221 2024-11-11T03:15:27,930 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:42221 connecting to ZooKeeper ensemble=127.0.0.1:54485 2024-11-11T03:15:27,961 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:422210x0, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-11T03:15:27,964 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:42221-0x101050601d00000 connected 2024-11-11T03:15:27,993 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:27,996 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:28,005 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:28,009 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939, hbase.cluster.distributed=false 2024-11-11T03:15:28,033 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-11T03:15:28,040 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=42221 2024-11-11T03:15:28,040 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=42221 2024-11-11T03:15:28,041 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=42221 2024-11-11T03:15:28,042 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=42221 2024-11-11T03:15:28,046 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=42221 2024-11-11T03:15:28,164 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/59e6b60f6143:0 server-side Connection retries=45 2024-11-11T03:15:28,166 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,167 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,167 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-11T03:15:28,167 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,167 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-11T03:15:28,171 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-11T03:15:28,173 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-11T03:15:28,174 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36869 2024-11-11T03:15:28,176 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36869 connecting to ZooKeeper ensemble=127.0.0.1:54485 2024-11-11T03:15:28,177 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:28,182 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:28,189 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:368690x0, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-11T03:15:28,190 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36869-0x101050601d00001 connected 2024-11-11T03:15:28,190 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:28,194 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-11T03:15:28,203 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-11T03:15:28,206 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-11T03:15:28,211 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-11T03:15:28,212 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36869 2024-11-11T03:15:28,213 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36869 2024-11-11T03:15:28,216 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36869 2024-11-11T03:15:28,217 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36869 2024-11-11T03:15:28,217 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36869 2024-11-11T03:15:28,237 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/59e6b60f6143:0 server-side Connection retries=45 2024-11-11T03:15:28,237 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,237 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,238 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-11T03:15:28,238 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,238 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-11T03:15:28,238 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-11T03:15:28,239 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-11T03:15:28,240 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:45545 2024-11-11T03:15:28,242 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:45545 connecting to ZooKeeper ensemble=127.0.0.1:54485 2024-11-11T03:15:28,244 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:28,248 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:28,256 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:455450x0, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-11T03:15:28,257 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45545-0x101050601d00002 connected 2024-11-11T03:15:28,257 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:28,258 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-11T03:15:28,260 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-11T03:15:28,261 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-11T03:15:28,263 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-11T03:15:28,264 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45545 2024-11-11T03:15:28,266 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45545 2024-11-11T03:15:28,267 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45545 2024-11-11T03:15:28,269 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45545 2024-11-11T03:15:28,269 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45545 2024-11-11T03:15:28,286 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/59e6b60f6143:0 server-side Connection retries=45 2024-11-11T03:15:28,287 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,287 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,287 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-11T03:15:28,287 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:28,287 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-11T03:15:28,287 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-11T03:15:28,287 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-11T03:15:28,288 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41395 2024-11-11T03:15:28,290 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:41395 connecting to ZooKeeper ensemble=127.0.0.1:54485 2024-11-11T03:15:28,291 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:28,293 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:28,303 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:413950x0, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-11T03:15:28,304 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:413950x0, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:28,304 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:41395-0x101050601d00003 connected 2024-11-11T03:15:28,304 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-11T03:15:28,305 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-11T03:15:28,306 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-11T03:15:28,308 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-11T03:15:28,309 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41395 2024-11-11T03:15:28,309 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41395 2024-11-11T03:15:28,310 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41395 2024-11-11T03:15:28,310 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41395 2024-11-11T03:15:28,311 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41395 2024-11-11T03:15:28,327 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;59e6b60f6143:42221 2024-11-11T03:15:28,328 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/59e6b60f6143,42221,1731294927435 2024-11-11T03:15:28,335 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:28,335 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:28,335 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:28,335 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:28,337 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/59e6b60f6143,42221,1731294927435 2024-11-11T03:15:28,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-11T03:15:28,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-11T03:15:28,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:28,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-11T03:15:28,358 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:28,358 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:28,358 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:28,359 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-11-11T03:15:28,360 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/59e6b60f6143,42221,1731294927435 from backup master directory 2024-11-11T03:15:28,363 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/59e6b60f6143,42221,1731294927435 2024-11-11T03:15:28,363 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:28,363 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:28,363 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:28,363 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:28,364 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-11T03:15:28,364 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=59e6b60f6143,42221,1731294927435 2024-11-11T03:15:28,366 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-11-11T03:15:28,367 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-11-11T03:15:28,430 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/hbase.id] with ID: 673c34d3-1e4b-47a4-a782-d219e3512096 2024-11-11T03:15:28,430 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/.tmp/hbase.id 2024-11-11T03:15:28,437 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:28,437 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:28,441 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:45600 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:34821:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45600 dst: /127.0.0.1:34821 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:28,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775776_1004 (size=42) 2024-11-11T03:15:28,447 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:28,447 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/.tmp/hbase.id]:[hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/hbase.id] 2024-11-11T03:15:28,490 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:28,494 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-11-11T03:15:28,513 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 17ms. 2024-11-11T03:15:28,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:28,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:28,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:28,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:28,529 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:28,529 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:28,532 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:45626 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:34821:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45626 dst: /127.0.0.1:34821 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:28,538 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775760_1006 (size=196) 2024-11-11T03:15:28,539 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:28,555 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-11T03:15:28,557 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-11-11T03:15:28,563 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-11T03:15:28,591 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:28,591 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:28,594 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:45636 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:34821:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45636 dst: /127.0.0.1:34821 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:28,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775744_1008 (size=1189) 2024-11-11T03:15:28,600 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:28,618 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store 2024-11-11T03:15:28,633 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:28,633 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:28,639 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:44798 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:44249:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:44798 dst: /127.0.0.1:44249 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:28,643 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_-9223372036854775728_1010 (size=34) 2024-11-11T03:15:28,644 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:28,649 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-11-11T03:15:28,651 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:28,653 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-11T03:15:28,653 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:28,653 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:28,654 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-11T03:15:28,654 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:28,655 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:28,656 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731294928653Disabling compacts and flushes for region at 1731294928653Disabling writes for close at 1731294928654 (+1 ms)Writing region close event to WAL at 1731294928655 (+1 ms)Closed at 1731294928655 2024-11-11T03:15:28,657 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/.initializing 2024-11-11T03:15:28,658 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/WALs/59e6b60f6143,42221,1731294927435 2024-11-11T03:15:28,666 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-11T03:15:28,679 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C42221%2C1731294927435, suffix=, logDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/WALs/59e6b60f6143,42221,1731294927435, archiveDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/oldWALs, maxLogs=10 2024-11-11T03:15:28,708 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/WALs/59e6b60f6143,42221,1731294927435/59e6b60f6143%2C42221%2C1731294927435.1731294928683, exclude list is [], retry=0 2024-11-11T03:15:28,726 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:414) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:473) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:468) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:28,728 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44249,DS-713ba5f4-b3a0-49fe-a9cd-2acddc4f9ee6,DISK] 2024-11-11T03:15:28,728 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45271,DS-28348765-dcc6-4606-a18e-329f4b7e00f5,DISK] 2024-11-11T03:15:28,728 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34821,DS-11f7bbb6-ae48-466e-9dc3-94891c1bd02e,DISK] 2024-11-11T03:15:28,731 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-11-11T03:15:28,775 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/WALs/59e6b60f6143,42221,1731294927435/59e6b60f6143%2C42221%2C1731294927435.1731294928683 2024-11-11T03:15:28,776 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:37915:37915),(127.0.0.1/127.0.0.1:36633:36633),(127.0.0.1/127.0.0.1:45315:45315)] 2024-11-11T03:15:28,777 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-11-11T03:15:28,777 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:28,780 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,781 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,828 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,858 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-11-11T03:15:28,862 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:28,864 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:28,865 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,868 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-11-11T03:15:28,869 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:28,870 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-11T03:15:28,870 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,872 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-11-11T03:15:28,873 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:28,874 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-11T03:15:28,874 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,877 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-11-11T03:15:28,877 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:28,878 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-11T03:15:28,879 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,882 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,883 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,888 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,889 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,892 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-11T03:15:28,895 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:28,901 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-11T03:15:28,902 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68879524, jitterRate=0.02638489007949829}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-11T03:15:28,908 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1731294928794Initializing all the Stores at 1731294928797 (+3 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294928797Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294928797Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294928798 (+1 ms)Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294928798Cleaning up temporary data from old regions at 1731294928889 (+91 ms)Region opened successfully at 1731294928908 (+19 ms) 2024-11-11T03:15:28,909 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-11-11T03:15:28,942 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@18729a8a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=59e6b60f6143/172.17.0.2:0 2024-11-11T03:15:28,974 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-11-11T03:15:28,985 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-11-11T03:15:28,985 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-11-11T03:15:28,988 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-11-11T03:15:28,989 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-11-11T03:15:28,994 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 4 msec 2024-11-11T03:15:28,994 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-11-11T03:15:29,019 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-11-11T03:15:29,028 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-11-11T03:15:29,031 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-11-11T03:15:29,034 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-11-11T03:15:29,035 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-11-11T03:15:29,037 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-11-11T03:15:29,039 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-11-11T03:15:29,043 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-11-11T03:15:29,044 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-11-11T03:15:29,045 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-11-11T03:15:29,048 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-11-11T03:15:29,065 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-11-11T03:15:29,067 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-11-11T03:15:29,071 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:29,071 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:29,071 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:29,071 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:29,071 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,071 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,071 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,071 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,074 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=59e6b60f6143,42221,1731294927435, sessionid=0x101050601d00000, setting cluster-up flag (Was=false) 2024-11-11T03:15:29,087 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,087 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,087 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,087 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,093 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-11-11T03:15:29,095 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=59e6b60f6143,42221,1731294927435 2024-11-11T03:15:29,101 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,101 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,101 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,101 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,108 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-11-11T03:15:29,109 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=59e6b60f6143,42221,1731294927435 2024-11-11T03:15:29,115 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-11-11T03:15:29,186 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-11-11T03:15:29,195 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-11-11T03:15:29,201 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-11-11T03:15:29,207 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 59e6b60f6143,42221,1731294927435 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-11-11T03:15:29,213 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/59e6b60f6143:0, corePoolSize=5, maxPoolSize=5 2024-11-11T03:15:29,213 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/59e6b60f6143:0, corePoolSize=5, maxPoolSize=5 2024-11-11T03:15:29,213 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/59e6b60f6143:0, corePoolSize=5, maxPoolSize=5 2024-11-11T03:15:29,213 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/59e6b60f6143:0, corePoolSize=5, maxPoolSize=5 2024-11-11T03:15:29,214 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/59e6b60f6143:0, corePoolSize=10, maxPoolSize=10 2024-11-11T03:15:29,214 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,214 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/59e6b60f6143:0, corePoolSize=2, maxPoolSize=2 2024-11-11T03:15:29,214 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,214 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(746): ClusterId : 673c34d3-1e4b-47a4-a782-d219e3512096 2024-11-11T03:15:29,214 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(746): ClusterId : 673c34d3-1e4b-47a4-a782-d219e3512096 2024-11-11T03:15:29,215 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(746): ClusterId : 673c34d3-1e4b-47a4-a782-d219e3512096 2024-11-11T03:15:29,218 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-11T03:15:29,218 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-11T03:15:29,218 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-11T03:15:29,219 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1731294959219 2024-11-11T03:15:29,220 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-11T03:15:29,220 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-11-11T03:15:29,222 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-11-11T03:15:29,223 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-11-11T03:15:29,226 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-11T03:15:29,226 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-11T03:15:29,226 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-11T03:15:29,226 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-11T03:15:29,226 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-11T03:15:29,226 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-11T03:15:29,228 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-11-11T03:15:29,228 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:29,228 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-11-11T03:15:29,229 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-11-11T03:15:29,229 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-11-11T03:15:29,228 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-11-11T03:15:29,231 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-11T03:15:29,231 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-11T03:15:29,231 DEBUG [RS:1;59e6b60f6143:45545 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4c0a0c0c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=59e6b60f6143/172.17.0.2:0 2024-11-11T03:15:29,232 DEBUG [RS:0;59e6b60f6143:36869 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@502a7c77, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=59e6b60f6143/172.17.0.2:0 2024-11-11T03:15:29,230 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,234 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-11T03:15:29,235 DEBUG [RS:2;59e6b60f6143:41395 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@244f5bdf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=59e6b60f6143/172.17.0.2:0 2024-11-11T03:15:29,239 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-11-11T03:15:29,240 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-11-11T03:15:29,241 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-11-11T03:15:29,247 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-11-11T03:15:29,247 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-11-11T03:15:29,250 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:29,250 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:29,256 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.large.0-1731294929249,5,FailOnTimeoutGroup] 2024-11-11T03:15:29,257 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.small.0-1731294929257,5,FailOnTimeoutGroup] 2024-11-11T03:15:29,257 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,258 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-11-11T03:15:29,258 DEBUG [RS:2;59e6b60f6143:41395 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;59e6b60f6143:41395 2024-11-11T03:15:29,258 DEBUG [RS:1;59e6b60f6143:45545 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;59e6b60f6143:45545 2024-11-11T03:15:29,259 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,259 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,261 DEBUG [RS:0;59e6b60f6143:36869 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;59e6b60f6143:36869 2024-11-11T03:15:29,262 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-11T03:15:29,262 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-11T03:15:29,262 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-11T03:15:29,262 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-11T03:15:29,262 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-11T03:15:29,262 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-11T03:15:29,262 DEBUG [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-11T03:15:29,262 DEBUG [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-11T03:15:29,262 DEBUG [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-11T03:15:29,265 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(2659): reportForDuty to master=59e6b60f6143,42221,1731294927435 with port=45545, startcode=1731294928236 2024-11-11T03:15:29,265 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(2659): reportForDuty to master=59e6b60f6143,42221,1731294927435 with port=36869, startcode=1731294928124 2024-11-11T03:15:29,265 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(2659): reportForDuty to master=59e6b60f6143,42221,1731294927435 with port=41395, startcode=1731294928286 2024-11-11T03:15:29,271 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:45658 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:34821:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45658 dst: /127.0.0.1:34821 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:29,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775788_1002 (size=7) 2024-11-11T03:15:29,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_-9223372036854775789_1002 (size=7) 2024-11-11T03:15:29,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775712_1013 (size=1321) 2024-11-11T03:15:29,279 DEBUG [RS:2;59e6b60f6143:41395 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-11T03:15:29,279 DEBUG [RS:1;59e6b60f6143:45545 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-11T03:15:29,279 DEBUG [RS:0;59e6b60f6143:36869 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-11T03:15:29,280 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:29,281 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-11-11T03:15:29,282 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939 2024-11-11T03:15:29,293 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:29,293 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:29,306 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:53060 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:45271:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53060 dst: /127.0.0.1:45271 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:29,314 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775696_1015 (size=32) 2024-11-11T03:15:29,317 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:29,319 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:29,324 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-11T03:15:29,327 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-11T03:15:29,327 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:29,327 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44113, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-11-11T03:15:29,327 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:47633, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-11-11T03:15:29,327 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54055, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-11-11T03:15:29,328 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:29,328 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-11T03:15:29,331 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-11T03:15:29,331 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:29,332 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:29,332 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-11T03:15:29,335 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42221 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 59e6b60f6143,45545,1731294928236 2024-11-11T03:15:29,335 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-11T03:15:29,335 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:29,336 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:29,337 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-11T03:15:29,338 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42221 {}] master.ServerManager(517): Registering regionserver=59e6b60f6143,45545,1731294928236 2024-11-11T03:15:29,339 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-11T03:15:29,339 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:29,340 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:29,341 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-11T03:15:29,342 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740 2024-11-11T03:15:29,343 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740 2024-11-11T03:15:29,347 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-11T03:15:29,347 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-11T03:15:29,348 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-11T03:15:29,350 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42221 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 59e6b60f6143,41395,1731294928286 2024-11-11T03:15:29,351 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42221 {}] master.ServerManager(517): Registering regionserver=59e6b60f6143,41395,1731294928286 2024-11-11T03:15:29,354 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-11T03:15:29,355 DEBUG [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939 2024-11-11T03:15:29,355 DEBUG [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:44063 2024-11-11T03:15:29,355 DEBUG [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-11T03:15:29,356 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42221 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 59e6b60f6143,36869,1731294928124 2024-11-11T03:15:29,357 DEBUG [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939 2024-11-11T03:15:29,357 DEBUG [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:44063 2024-11-11T03:15:29,357 DEBUG [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-11T03:15:29,357 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42221 {}] master.ServerManager(517): Registering regionserver=59e6b60f6143,36869,1731294928124 2024-11-11T03:15:29,361 DEBUG [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939 2024-11-11T03:15:29,362 DEBUG [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:44063 2024-11-11T03:15:29,362 DEBUG [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-11T03:15:29,364 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-11T03:15:29,365 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-11T03:15:29,365 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66466480, jitterRate=-0.009572267532348633}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-11T03:15:29,368 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1731294929319Initializing all the Stores at 1731294929323 (+4 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294929324 (+1 ms)Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294929324Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294929324Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294929324Cleaning up temporary data from old regions at 1731294929347 (+23 ms)Region opened successfully at 1731294929368 (+21 ms) 2024-11-11T03:15:29,368 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-11T03:15:29,368 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-11T03:15:29,368 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-11T03:15:29,368 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-11T03:15:29,368 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-11T03:15:29,369 DEBUG [RS:1;59e6b60f6143:45545 {}] zookeeper.ZKUtil(111): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/59e6b60f6143,45545,1731294928236 2024-11-11T03:15:29,369 DEBUG [RS:2;59e6b60f6143:41395 {}] zookeeper.ZKUtil(111): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/59e6b60f6143,41395,1731294928286 2024-11-11T03:15:29,369 WARN [RS:2;59e6b60f6143:41395 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-11T03:15:29,369 WARN [RS:1;59e6b60f6143:45545 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-11T03:15:29,369 INFO [RS:1;59e6b60f6143:45545 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-11T03:15:29,369 INFO [RS:2;59e6b60f6143:41395 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-11T03:15:29,369 DEBUG [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,45545,1731294928236 2024-11-11T03:15:29,369 DEBUG [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,41395,1731294928286 2024-11-11T03:15:29,370 DEBUG [RS:0;59e6b60f6143:36869 {}] zookeeper.ZKUtil(111): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/59e6b60f6143,36869,1731294928124 2024-11-11T03:15:29,370 WARN [RS:0;59e6b60f6143:36869 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-11T03:15:29,370 INFO [RS:0;59e6b60f6143:36869 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-11T03:15:29,370 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-11T03:15:29,370 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731294929368Disabling compacts and flushes for region at 1731294929368Disabling writes for close at 1731294929368Writing region close event to WAL at 1731294929370 (+2 ms)Closed at 1731294929370 2024-11-11T03:15:29,370 DEBUG [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,36869,1731294928124 2024-11-11T03:15:29,372 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [59e6b60f6143,45545,1731294928236] 2024-11-11T03:15:29,372 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [59e6b60f6143,41395,1731294928286] 2024-11-11T03:15:29,372 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [59e6b60f6143,36869,1731294928124] 2024-11-11T03:15:29,374 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-11T03:15:29,374 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-11-11T03:15:29,383 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-11-11T03:15:29,393 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-11T03:15:29,398 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-11-11T03:15:29,403 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-11T03:15:29,403 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-11T03:15:29,403 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-11T03:15:29,420 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-11T03:15:29,420 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-11T03:15:29,420 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-11T03:15:29,430 INFO [RS:1;59e6b60f6143:45545 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-11T03:15:29,430 INFO [RS:0;59e6b60f6143:36869 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-11T03:15:29,430 INFO [RS:2;59e6b60f6143:41395 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-11T03:15:29,430 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,430 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,430 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,432 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-11T03:15:29,432 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-11T03:15:29,432 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-11T03:15:29,439 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-11T03:15:29,439 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-11T03:15:29,439 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-11T03:15:29,440 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,440 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,440 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,441 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,441 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,441 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,441 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,441 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,441 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,441 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,441 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/59e6b60f6143:0, corePoolSize=2, maxPoolSize=2 2024-11-11T03:15:29,442 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/59e6b60f6143:0, corePoolSize=2, maxPoolSize=2 2024-11-11T03:15:29,442 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,442 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,443 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,443 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,443 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,443 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,443 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,443 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:29,443 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:29,443 DEBUG [RS:1;59e6b60f6143:45545 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:29,443 DEBUG [RS:2;59e6b60f6143:41395 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:29,443 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,443 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,443 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,444 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,444 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,444 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/59e6b60f6143:0, corePoolSize=2, maxPoolSize=2 2024-11-11T03:15:29,444 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,444 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,444 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,444 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,444 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,445 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:29,445 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:29,445 DEBUG [RS:0;59e6b60f6143:36869 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:29,445 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,445 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,445 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,445 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,445 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,446 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,446 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,446 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,45545,1731294928236-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-11T03:15:29,446 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,446 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,446 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,446 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41395,1731294928286-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-11T03:15:29,448 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,448 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,448 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,448 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,448 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,449 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,36869,1731294928124-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-11T03:15:29,469 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-11T03:15:29,471 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,45545,1731294928236-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,471 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,472 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.Replication(171): 59e6b60f6143,45545,1731294928236 started 2024-11-11T03:15:29,475 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-11T03:15:29,475 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41395,1731294928286-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,475 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-11T03:15:29,475 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,475 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,36869,1731294928124-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,475 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.Replication(171): 59e6b60f6143,41395,1731294928286 started 2024-11-11T03:15:29,475 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,475 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.Replication(171): 59e6b60f6143,36869,1731294928124 started 2024-11-11T03:15:29,492 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,492 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(1482): Serving as 59e6b60f6143,45545,1731294928236, RpcServer on 59e6b60f6143/172.17.0.2:45545, sessionid=0x101050601d00002 2024-11-11T03:15:29,493 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-11T03:15:29,493 DEBUG [RS:1;59e6b60f6143:45545 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 59e6b60f6143,45545,1731294928236 2024-11-11T03:15:29,493 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,45545,1731294928236' 2024-11-11T03:15:29,493 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-11T03:15:29,494 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-11T03:15:29,495 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-11T03:15:29,495 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-11T03:15:29,495 DEBUG [RS:1;59e6b60f6143:45545 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 59e6b60f6143,45545,1731294928236 2024-11-11T03:15:29,495 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,45545,1731294928236' 2024-11-11T03:15:29,495 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-11T03:15:29,496 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-11T03:15:29,496 DEBUG [RS:1;59e6b60f6143:45545 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-11T03:15:29,497 INFO [RS:1;59e6b60f6143:45545 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-11T03:15:29,497 INFO [RS:1;59e6b60f6143:45545 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-11T03:15:29,502 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,502 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:29,502 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1482): Serving as 59e6b60f6143,36869,1731294928124, RpcServer on 59e6b60f6143/172.17.0.2:36869, sessionid=0x101050601d00001 2024-11-11T03:15:29,502 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(1482): Serving as 59e6b60f6143,41395,1731294928286, RpcServer on 59e6b60f6143/172.17.0.2:41395, sessionid=0x101050601d00003 2024-11-11T03:15:29,502 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-11T03:15:29,502 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-11T03:15:29,502 DEBUG [RS:0;59e6b60f6143:36869 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 59e6b60f6143,36869,1731294928124 2024-11-11T03:15:29,502 DEBUG [RS:2;59e6b60f6143:41395 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 59e6b60f6143,41395,1731294928286 2024-11-11T03:15:29,503 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,36869,1731294928124' 2024-11-11T03:15:29,503 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,41395,1731294928286' 2024-11-11T03:15:29,503 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-11T03:15:29,503 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-11T03:15:29,503 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-11T03:15:29,503 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-11T03:15:29,504 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-11T03:15:29,504 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-11T03:15:29,504 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-11T03:15:29,504 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-11T03:15:29,504 DEBUG [RS:2;59e6b60f6143:41395 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 59e6b60f6143,41395,1731294928286 2024-11-11T03:15:29,504 DEBUG [RS:0;59e6b60f6143:36869 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 59e6b60f6143,36869,1731294928124 2024-11-11T03:15:29,504 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,41395,1731294928286' 2024-11-11T03:15:29,505 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,36869,1731294928124' 2024-11-11T03:15:29,505 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-11T03:15:29,505 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-11T03:15:29,505 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-11T03:15:29,505 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-11T03:15:29,506 DEBUG [RS:0;59e6b60f6143:36869 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-11T03:15:29,506 INFO [RS:0;59e6b60f6143:36869 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-11T03:15:29,506 DEBUG [RS:2;59e6b60f6143:41395 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-11T03:15:29,506 INFO [RS:0;59e6b60f6143:36869 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-11T03:15:29,506 INFO [RS:2;59e6b60f6143:41395 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-11T03:15:29,506 INFO [RS:2;59e6b60f6143:41395 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-11T03:15:29,549 WARN [59e6b60f6143:42221 {}] assignment.AssignmentManager(2443): No servers available; cannot place 1 unassigned regions. 2024-11-11T03:15:29,602 INFO [RS:1;59e6b60f6143:45545 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-11T03:15:29,606 INFO [RS:1;59e6b60f6143:45545 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C45545%2C1731294928236, suffix=, logDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,45545,1731294928236, archiveDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/oldWALs, maxLogs=32 2024-11-11T03:15:29,607 INFO [RS:0;59e6b60f6143:36869 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-11T03:15:29,607 INFO [RS:2;59e6b60f6143:41395 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-11T03:15:29,610 INFO [RS:0;59e6b60f6143:36869 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C36869%2C1731294928124, suffix=, logDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,36869,1731294928124, archiveDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/oldWALs, maxLogs=32 2024-11-11T03:15:29,610 INFO [RS:2;59e6b60f6143:41395 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C41395%2C1731294928286, suffix=, logDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,41395,1731294928286, archiveDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/oldWALs, maxLogs=32 2024-11-11T03:15:29,632 DEBUG [RS:1;59e6b60f6143:45545 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,45545,1731294928236/59e6b60f6143%2C45545%2C1731294928236.1731294929609, exclude list is [], retry=0 2024-11-11T03:15:29,634 DEBUG [RS:2;59e6b60f6143:41395 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,41395,1731294928286/59e6b60f6143%2C41395%2C1731294928286.1731294929614, exclude list is [], retry=0 2024-11-11T03:15:29,635 DEBUG [RS:0;59e6b60f6143:36869 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,36869,1731294928124/59e6b60f6143%2C36869%2C1731294928124.1731294929613, exclude list is [], retry=0 2024-11-11T03:15:29,639 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45271,DS-28348765-dcc6-4606-a18e-329f4b7e00f5,DISK] 2024-11-11T03:15:29,639 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34821,DS-11f7bbb6-ae48-466e-9dc3-94891c1bd02e,DISK] 2024-11-11T03:15:29,669 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44249,DS-713ba5f4-b3a0-49fe-a9cd-2acddc4f9ee6,DISK] 2024-11-11T03:15:29,670 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45271,DS-28348765-dcc6-4606-a18e-329f4b7e00f5,DISK] 2024-11-11T03:15:29,670 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45271,DS-28348765-dcc6-4606-a18e-329f4b7e00f5,DISK] 2024-11-11T03:15:29,670 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44249,DS-713ba5f4-b3a0-49fe-a9cd-2acddc4f9ee6,DISK] 2024-11-11T03:15:29,671 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34821,DS-11f7bbb6-ae48-466e-9dc3-94891c1bd02e,DISK] 2024-11-11T03:15:29,671 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34821,DS-11f7bbb6-ae48-466e-9dc3-94891c1bd02e,DISK] 2024-11-11T03:15:29,671 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44249,DS-713ba5f4-b3a0-49fe-a9cd-2acddc4f9ee6,DISK] 2024-11-11T03:15:29,676 INFO [RS:1;59e6b60f6143:45545 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,45545,1731294928236/59e6b60f6143%2C45545%2C1731294928236.1731294929609 2024-11-11T03:15:29,676 INFO [RS:2;59e6b60f6143:41395 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,41395,1731294928286/59e6b60f6143%2C41395%2C1731294928286.1731294929614 2024-11-11T03:15:29,676 INFO [RS:0;59e6b60f6143:36869 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,36869,1731294928124/59e6b60f6143%2C36869%2C1731294928124.1731294929613 2024-11-11T03:15:29,676 DEBUG [RS:1;59e6b60f6143:45545 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36633:36633),(127.0.0.1/127.0.0.1:45315:45315),(127.0.0.1/127.0.0.1:37915:37915)] 2024-11-11T03:15:29,677 DEBUG [RS:2;59e6b60f6143:41395 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36633:36633),(127.0.0.1/127.0.0.1:37915:37915),(127.0.0.1/127.0.0.1:45315:45315)] 2024-11-11T03:15:29,678 DEBUG [RS:0;59e6b60f6143:36869 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36633:36633),(127.0.0.1/127.0.0.1:37915:37915),(127.0.0.1/127.0.0.1:45315:45315)] 2024-11-11T03:15:29,802 DEBUG [59e6b60f6143:42221 {}] assignment.AssignmentManager(2464): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-11-11T03:15:29,810 DEBUG [59e6b60f6143:42221 {}] balancer.BalancerClusterState(204): Hosts are {59e6b60f6143=0} racks are {/default-rack=0} 2024-11-11T03:15:29,816 DEBUG [59e6b60f6143:42221 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-11T03:15:29,817 DEBUG [59e6b60f6143:42221 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-11T03:15:29,817 DEBUG [59e6b60f6143:42221 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-11T03:15:29,817 DEBUG [59e6b60f6143:42221 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-11T03:15:29,817 DEBUG [59e6b60f6143:42221 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-11T03:15:29,817 DEBUG [59e6b60f6143:42221 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-11T03:15:29,817 INFO [59e6b60f6143:42221 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-11T03:15:29,817 INFO [59e6b60f6143:42221 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-11T03:15:29,817 INFO [59e6b60f6143:42221 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-11T03:15:29,817 DEBUG [59e6b60f6143:42221 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-11T03:15:29,825 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=59e6b60f6143,36869,1731294928124 2024-11-11T03:15:29,832 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 59e6b60f6143,36869,1731294928124, state=OPENING 2024-11-11T03:15:29,838 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-11-11T03:15:29,840 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,840 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,840 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,840 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:29,841 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:29,841 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:29,841 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:29,841 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:29,842 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-11T03:15:29,844 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=59e6b60f6143,36869,1731294928124}] 2024-11-11T03:15:30,020 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-11T03:15:30,023 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46287, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-11T03:15:30,036 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-11-11T03:15:30,036 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-11T03:15:30,037 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-11-11T03:15:30,040 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C36869%2C1731294928124.meta, suffix=.meta, logDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,36869,1731294928124, archiveDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/oldWALs, maxLogs=32 2024-11-11T03:15:30,057 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,36869,1731294928124/59e6b60f6143%2C36869%2C1731294928124.meta.1731294930042.meta, exclude list is [], retry=0 2024-11-11T03:15:30,061 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45271,DS-28348765-dcc6-4606-a18e-329f4b7e00f5,DISK] 2024-11-11T03:15:30,062 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44249,DS-713ba5f4-b3a0-49fe-a9cd-2acddc4f9ee6,DISK] 2024-11-11T03:15:30,062 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34821,DS-11f7bbb6-ae48-466e-9dc3-94891c1bd02e,DISK] 2024-11-11T03:15:30,065 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/WALs/59e6b60f6143,36869,1731294928124/59e6b60f6143%2C36869%2C1731294928124.meta.1731294930042.meta 2024-11-11T03:15:30,065 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36633:36633),(127.0.0.1/127.0.0.1:37915:37915),(127.0.0.1/127.0.0.1:45315:45315)] 2024-11-11T03:15:30,066 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-11-11T03:15:30,067 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-11-11T03:15:30,070 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-11-11T03:15:30,075 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-11-11T03:15:30,079 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-11-11T03:15:30,080 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:30,080 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-11-11T03:15:30,080 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-11-11T03:15:30,083 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-11T03:15:30,085 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-11T03:15:30,085 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:30,086 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:30,086 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-11T03:15:30,087 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-11T03:15:30,088 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:30,088 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:30,088 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-11T03:15:30,090 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-11T03:15:30,090 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:30,090 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:30,091 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-11T03:15:30,092 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-11T03:15:30,092 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:30,093 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:30,093 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-11T03:15:30,094 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740 2024-11-11T03:15:30,097 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740 2024-11-11T03:15:30,100 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-11T03:15:30,100 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-11T03:15:30,101 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-11T03:15:30,104 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-11T03:15:30,106 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73538836, jitterRate=0.095814049243927}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-11T03:15:30,106 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-11-11T03:15:30,108 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1731294930081Writing region info on filesystem at 1731294930081Initializing all the Stores at 1731294930083 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294930083Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294930083Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294930083Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294930083Cleaning up temporary data from old regions at 1731294930100 (+17 ms)Running coprocessor post-open hooks at 1731294930106 (+6 ms)Region opened successfully at 1731294930108 (+2 ms) 2024-11-11T03:15:30,116 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1731294930011 2024-11-11T03:15:30,127 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-11-11T03:15:30,128 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-11-11T03:15:30,129 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=59e6b60f6143,36869,1731294928124 2024-11-11T03:15:30,132 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 59e6b60f6143,36869,1731294928124, state=OPEN 2024-11-11T03:15:30,134 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-11T03:15:30,134 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-11T03:15:30,134 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-11T03:15:30,134 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-11T03:15:30,134 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:30,134 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:30,134 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:30,134 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:30,135 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=59e6b60f6143,36869,1731294928124 2024-11-11T03:15:30,140 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-11-11T03:15:30,141 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=59e6b60f6143,36869,1731294928124 in 291 msec 2024-11-11T03:15:30,147 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-11-11T03:15:30,147 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 760 msec 2024-11-11T03:15:30,148 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-11T03:15:30,148 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-11-11T03:15:30,170 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-11T03:15:30,171 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=59e6b60f6143,36869,1731294928124, seqNum=-1] 2024-11-11T03:15:30,193 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-11T03:15:30,195 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33497, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-11T03:15:30,250 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1050 sec 2024-11-11T03:15:30,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1731294930250, completionTime=-1 2024-11-11T03:15:30,253 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-11-11T03:15:30,253 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] assignment.AssignmentManager(1756): Joining cluster... 2024-11-11T03:15:30,287 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] assignment.AssignmentManager(1768): Number of RegionServers=3 2024-11-11T03:15:30,287 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1731294990287 2024-11-11T03:15:30,287 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1731295050287 2024-11-11T03:15:30,287 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] assignment.AssignmentManager(1775): Joined the cluster in 33 msec 2024-11-11T03:15:30,289 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-11-11T03:15:30,295 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,42221,1731294927435-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:30,295 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,42221,1731294927435-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:30,295 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,42221,1731294927435-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:30,297 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-59e6b60f6143:42221, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:30,297 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:30,298 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:30,303 DEBUG [master/59e6b60f6143:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-11-11T03:15:30,325 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.961sec 2024-11-11T03:15:30,327 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-11-11T03:15:30,328 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-11-11T03:15:30,329 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-11-11T03:15:30,329 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-11-11T03:15:30,330 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-11-11T03:15:30,330 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,42221,1731294927435-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-11T03:15:30,331 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,42221,1731294927435-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-11-11T03:15:30,336 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-11-11T03:15:30,337 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-11-11T03:15:30,337 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,42221,1731294927435-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:30,343 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@65a66577, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-11T03:15:30,348 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-11-11T03:15:30,348 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-11-11T03:15:30,352 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 59e6b60f6143,42221,-1 for getting cluster id 2024-11-11T03:15:30,355 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-11-11T03:15:30,363 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '673c34d3-1e4b-47a4-a782-d219e3512096' 2024-11-11T03:15:30,366 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-11-11T03:15:30,366 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "673c34d3-1e4b-47a4-a782-d219e3512096" 2024-11-11T03:15:30,366 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@504c202b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-11T03:15:30,366 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [59e6b60f6143,42221,-1] 2024-11-11T03:15:30,369 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-11-11T03:15:30,371 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:30,372 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42642, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-11-11T03:15:30,374 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6bf285b2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-11T03:15:30,375 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-11T03:15:30,382 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=59e6b60f6143,36869,1731294928124, seqNum=-1] 2024-11-11T03:15:30,382 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-11T03:15:30,384 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48160, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-11T03:15:30,404 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=59e6b60f6143,42221,1731294927435 2024-11-11T03:15:30,408 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-11-11T03:15:30,413 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 59e6b60f6143,42221,1731294927435 2024-11-11T03:15:30,416 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@5aa7f78b 2024-11-11T03:15:30,417 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-11-11T03:15:30,419 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42650, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-11-11T03:15:30,425 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-11T03:15:30,433 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-11-11T03:15:30,435 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-11-11T03:15:30,437 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-11-11T03:15:30,438 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:30,440 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-11T03:15:30,443 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-11T03:15:30,450 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:30,450 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:30,455 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:45686 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:34821:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45686 dst: /127.0.0.1:34821 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:30,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775680_1021 (size=392) 2024-11-11T03:15:30,462 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:30,465 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => fb960b26c6eb2ef1ad22da807871cd25, NAME => 'TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939 2024-11-11T03:15:30,470 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:30,471 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:30,473 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:53114 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:45271:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53114 dst: /127.0.0.1:45271 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:30,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775664_1023 (size=51) 2024-11-11T03:15:30,481 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:30,482 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:30,482 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing fb960b26c6eb2ef1ad22da807871cd25, disabling compactions & flushes 2024-11-11T03:15:30,482 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:30,482 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:30,482 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. after waiting 0 ms 2024-11-11T03:15:30,482 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:30,482 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:30,482 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for fb960b26c6eb2ef1ad22da807871cd25: Waiting for close lock at 1731294930482Disabling compacts and flushes for region at 1731294930482Disabling writes for close at 1731294930482Writing region close event to WAL at 1731294930482Closed at 1731294930482 2024-11-11T03:15:30,484 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-11-11T03:15:30,490 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1731294930485"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1731294930485"}]},"ts":"1731294930485"} 2024-11-11T03:15:30,495 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-11-11T03:15:30,497 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-11-11T03:15:30,500 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731294930497"}]},"ts":"1731294930497"} 2024-11-11T03:15:30,505 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-11-11T03:15:30,506 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {59e6b60f6143=0} racks are {/default-rack=0} 2024-11-11T03:15:30,508 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-11T03:15:30,508 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-11T03:15:30,508 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-11T03:15:30,508 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-11T03:15:30,508 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-11T03:15:30,508 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-11T03:15:30,508 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-11T03:15:30,508 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-11T03:15:30,508 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-11T03:15:30,508 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-11T03:15:30,510 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=fb960b26c6eb2ef1ad22da807871cd25, ASSIGN}] 2024-11-11T03:15:30,514 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=fb960b26c6eb2ef1ad22da807871cd25, ASSIGN 2024-11-11T03:15:30,517 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=fb960b26c6eb2ef1ad22da807871cd25, ASSIGN; state=OFFLINE, location=59e6b60f6143,36869,1731294928124; forceNewPlan=false, retain=false 2024-11-11T03:15:30,559 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-11T03:15:30,670 INFO [59e6b60f6143:42221 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-11-11T03:15:30,671 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=fb960b26c6eb2ef1ad22da807871cd25, regionState=OPENING, regionLocation=59e6b60f6143,36869,1731294928124 2024-11-11T03:15:30,675 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=fb960b26c6eb2ef1ad22da807871cd25, ASSIGN because future has completed 2024-11-11T03:15:30,676 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure fb960b26c6eb2ef1ad22da807871cd25, server=59e6b60f6143,36869,1731294928124}] 2024-11-11T03:15:30,769 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-11T03:15:30,836 INFO [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:30,837 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => fb960b26c6eb2ef1ad22da807871cd25, NAME => 'TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25.', STARTKEY => '', ENDKEY => ''} 2024-11-11T03:15:30,837 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,838 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:30,838 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,838 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,840 INFO [StoreOpener-fb960b26c6eb2ef1ad22da807871cd25-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,842 INFO [StoreOpener-fb960b26c6eb2ef1ad22da807871cd25-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region fb960b26c6eb2ef1ad22da807871cd25 columnFamilyName cf 2024-11-11T03:15:30,842 DEBUG [StoreOpener-fb960b26c6eb2ef1ad22da807871cd25-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:30,844 INFO [StoreOpener-fb960b26c6eb2ef1ad22da807871cd25-1 {}] regionserver.HStore(327): Store=fb960b26c6eb2ef1ad22da807871cd25/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-11T03:15:30,844 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,845 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,846 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,847 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,847 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,850 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,856 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-11T03:15:30,857 INFO [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened fb960b26c6eb2ef1ad22da807871cd25; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71261667, jitterRate=0.06188158690929413}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-11-11T03:15:30,857 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:30,859 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for fb960b26c6eb2ef1ad22da807871cd25: Running coprocessor pre-open hook at 1731294930838Writing region info on filesystem at 1731294930838Initializing all the Stores at 1731294930840 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294930840Cleaning up temporary data from old regions at 1731294930847 (+7 ms)Running coprocessor post-open hooks at 1731294930857 (+10 ms)Region opened successfully at 1731294930858 (+1 ms) 2024-11-11T03:15:30,861 INFO [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25., pid=6, masterSystemTime=1731294930829 2024-11-11T03:15:30,865 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:30,865 INFO [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:30,866 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=fb960b26c6eb2ef1ad22da807871cd25, regionState=OPEN, openSeqNum=2, regionLocation=59e6b60f6143,36869,1731294928124 2024-11-11T03:15:30,871 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure fb960b26c6eb2ef1ad22da807871cd25, server=59e6b60f6143,36869,1731294928124 because future has completed 2024-11-11T03:15:30,877 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-11-11T03:15:30,877 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure fb960b26c6eb2ef1ad22da807871cd25, server=59e6b60f6143,36869,1731294928124 in 198 msec 2024-11-11T03:15:30,882 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-11-11T03:15:30,882 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=fb960b26c6eb2ef1ad22da807871cd25, ASSIGN in 367 msec 2024-11-11T03:15:30,883 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-11-11T03:15:30,884 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731294930884"}]},"ts":"1731294930884"} 2024-11-11T03:15:30,887 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-11-11T03:15:30,888 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-11-11T03:15:30,891 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 461 msec 2024-11-11T03:15:31,079 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-11T03:15:31,080 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-11-11T03:15:31,080 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-11-11T03:15:31,081 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-11T03:15:31,086 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-11-11T03:15:31,086 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-11T03:15:31,087 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-11-11T03:15:31,095 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25., hostname=59e6b60f6143,36869,1731294928124, seqNum=2] 2024-11-11T03:15:31,104 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-11-11T03:15:31,109 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-11-11T03:15:31,111 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-11-11T03:15:31,111 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-11T03:15:31,112 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-11-11T03:15:31,113 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-11-11T03:15:31,219 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-11T03:15:31,276 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36869 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-11-11T03:15:31,277 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:31,283 INFO [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing fb960b26c6eb2ef1ad22da807871cd25 1/1 column families, dataSize=32 B heapSize=360 B 2024-11-11T03:15:31,341 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25/.tmp/cf/2bf164b48bd94e79a4abd1ec4d4ebb3e is 36, key is row/cf:cq/1731294931097/Put/seqid=0 2024-11-11T03:15:31,347 WARN [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:31,348 WARN [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:31,352 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1119780539_22 at /127.0.0.1:37406 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:34821:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37406 dst: /127.0.0.1:34821 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:31,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775648_1025 (size=4787) 2024-11-11T03:15:31,429 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-11T03:15:31,739 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-11T03:15:31,757 WARN [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:31,758 INFO [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25/.tmp/cf/2bf164b48bd94e79a4abd1ec4d4ebb3e 2024-11-11T03:15:31,809 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25/.tmp/cf/2bf164b48bd94e79a4abd1ec4d4ebb3e as hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25/cf/2bf164b48bd94e79a4abd1ec4d4ebb3e 2024-11-11T03:15:31,819 INFO [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25/cf/2bf164b48bd94e79a4abd1ec4d4ebb3e, entries=1, sequenceid=5, filesize=4.7 K 2024-11-11T03:15:31,828 INFO [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for fb960b26c6eb2ef1ad22da807871cd25 in 546ms, sequenceid=5, compaction requested=false 2024-11-11T03:15:31,829 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-11-11T03:15:31,831 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for fb960b26c6eb2ef1ad22da807871cd25: 2024-11-11T03:15:31,831 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:31,833 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-11-11T03:15:31,835 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-11-11T03:15:31,842 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-11-11T03:15:31,842 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 724 msec 2024-11-11T03:15:31,846 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 738 msec 2024-11-11T03:15:32,153 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775725_1010 (size=34) 2024-11-11T03:15:32,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775724_1010 (size=34) 2024-11-11T03:15:32,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775773_1004 (size=42) 2024-11-11T03:15:32,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_-9223372036854775772_1004 (size=42) 2024-11-11T03:15:32,220 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_-9223372036854775757_1006 (size=196) 2024-11-11T03:15:32,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775756_1006 (size=196) 2024-11-11T03:15:32,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775740_1008 (size=1189) 2024-11-11T03:15:32,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_-9223372036854775741_1008 (size=1189) 2024-11-11T03:15:32,248 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42221 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-11T03:15:32,249 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-11-11T03:15:32,262 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-11-11T03:15:32,262 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-11T03:15:32,263 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:32,267 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,268 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,268 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-11-11T03:15:32,268 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-11-11T03:15:32,268 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=98732422, stopped=false 2024-11-11T03:15:32,268 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=59e6b60f6143,42221,1731294927435 2024-11-11T03:15:32,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:32,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:32,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:32,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:32,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:32,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:32,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:32,270 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-11T03:15:32,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:32,271 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-11T03:15:32,271 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:32,271 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,271 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:32,271 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:32,272 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '59e6b60f6143,36869,1731294928124' ***** 2024-11-11T03:15:32,272 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-11T03:15:32,272 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '59e6b60f6143,45545,1731294928236' ***** 2024-11-11T03:15:32,272 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-11T03:15:32,272 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:32,272 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '59e6b60f6143,41395,1731294928286' ***** 2024-11-11T03:15:32,272 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-11T03:15:32,272 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-11T03:15:32,272 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-11T03:15:32,272 INFO [RS:1;59e6b60f6143:45545 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-11T03:15:32,273 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-11T03:15:32,273 INFO [RS:1;59e6b60f6143:45545 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-11T03:15:32,273 INFO [RS:0;59e6b60f6143:36869 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-11T03:15:32,273 INFO [RS:0;59e6b60f6143:36869 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-11T03:15:32,273 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(959): stopping server 59e6b60f6143,45545,1731294928236 2024-11-11T03:15:32,273 INFO [RS:1;59e6b60f6143:45545 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-11T03:15:32,273 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-11T03:15:32,273 INFO [RS:1;59e6b60f6143:45545 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;59e6b60f6143:45545. 2024-11-11T03:15:32,273 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(3091): Received CLOSE for fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:32,273 DEBUG [RS:1;59e6b60f6143:45545 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:32,273 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-11T03:15:32,273 DEBUG [RS:1;59e6b60f6143:45545 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,273 INFO [RS:2;59e6b60f6143:41395 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-11T03:15:32,273 INFO [RS:2;59e6b60f6143:41395 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-11T03:15:32,273 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(959): stopping server 59e6b60f6143,41395,1731294928286 2024-11-11T03:15:32,273 INFO [RS:2;59e6b60f6143:41395 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-11T03:15:32,273 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(976): stopping server 59e6b60f6143,45545,1731294928236; all regions closed. 2024-11-11T03:15:32,274 INFO [RS:2;59e6b60f6143:41395 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;59e6b60f6143:41395. 2024-11-11T03:15:32,274 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:32,274 DEBUG [RS:2;59e6b60f6143:41395 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:32,274 DEBUG [RS:2;59e6b60f6143:41395 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,274 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(976): stopping server 59e6b60f6143,41395,1731294928286; all regions closed. 2024-11-11T03:15:32,274 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(959): stopping server 59e6b60f6143,36869,1731294928124 2024-11-11T03:15:32,274 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-11T03:15:32,274 INFO [RS:0;59e6b60f6143:36869 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-11T03:15:32,274 INFO [RS:0;59e6b60f6143:36869 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;59e6b60f6143:36869. 2024-11-11T03:15:32,274 DEBUG [RS:0;59e6b60f6143:36869 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:32,274 DEBUG [RS:0;59e6b60f6143:36869 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,275 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing fb960b26c6eb2ef1ad22da807871cd25, disabling compactions & flushes 2024-11-11T03:15:32,275 INFO [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:32,275 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:32,275 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. after waiting 0 ms 2024-11-11T03:15:32,275 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-11T03:15:32,275 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:32,275 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-11T03:15:32,275 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-11T03:15:32,275 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-11-11T03:15:32,276 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-11-11T03:15:32,277 DEBUG [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1325): Online Regions={fb960b26c6eb2ef1ad22da807871cd25=TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25., 1588230740=hbase:meta,,1.1588230740} 2024-11-11T03:15:32,277 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-11T03:15:32,277 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-11T03:15:32,277 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-11T03:15:32,277 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-11T03:15:32,277 DEBUG [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, fb960b26c6eb2ef1ad22da807871cd25 2024-11-11T03:15:32,277 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-11T03:15:32,277 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-11-11T03:15:32,280 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_1073741827_1017 (size=93) 2024-11-11T03:15:32,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_1073741827_1017 (size=93) 2024-11-11T03:15:32,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_1073741827_1017 (size=93) 2024-11-11T03:15:32,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_1073741826_1016 (size=93) 2024-11-11T03:15:32,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_1073741826_1016 (size=93) 2024-11-11T03:15:32,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_1073741826_1016 (size=93) 2024-11-11T03:15:32,288 DEBUG [RS:2;59e6b60f6143:41395 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/oldWALs 2024-11-11T03:15:32,289 DEBUG [RS:1;59e6b60f6143:45545 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/oldWALs 2024-11-11T03:15:32,289 INFO [RS:2;59e6b60f6143:41395 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 59e6b60f6143%2C41395%2C1731294928286:(num 1731294929614) 2024-11-11T03:15:32,289 INFO [RS:1;59e6b60f6143:45545 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 59e6b60f6143%2C45545%2C1731294928236:(num 1731294929609) 2024-11-11T03:15:32,289 DEBUG [RS:2;59e6b60f6143:41395 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,289 DEBUG [RS:1;59e6b60f6143:45545 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,289 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:32,289 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:32,289 INFO [RS:2;59e6b60f6143:41395 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-11T03:15:32,289 INFO [RS:1;59e6b60f6143:45545 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-11T03:15:32,289 INFO [RS:1;59e6b60f6143:45545 {}] hbase.ChoreService(370): Chore service for: regionserver/59e6b60f6143:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-11T03:15:32,289 INFO [RS:2;59e6b60f6143:41395 {}] hbase.ChoreService(370): Chore service for: regionserver/59e6b60f6143:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-11T03:15:32,290 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-11T03:15:32,290 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-11T03:15:32,290 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-11T03:15:32,290 INFO [regionserver/59e6b60f6143:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-11T03:15:32,290 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-11T03:15:32,290 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-11T03:15:32,290 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-11T03:15:32,290 INFO [RS:1;59e6b60f6143:45545 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-11T03:15:32,290 INFO [RS:2;59e6b60f6143:41395 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-11T03:15:32,290 INFO [regionserver/59e6b60f6143:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-11T03:15:32,290 INFO [RS:1;59e6b60f6143:45545 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:45545 2024-11-11T03:15:32,290 INFO [RS:2;59e6b60f6143:41395 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41395 2024-11-11T03:15:32,295 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-11T03:15:32,295 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/59e6b60f6143,41395,1731294928286 2024-11-11T03:15:32,296 INFO [RS:2;59e6b60f6143:41395 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-11T03:15:32,297 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/59e6b60f6143,45545,1731294928236 2024-11-11T03:15:32,298 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [59e6b60f6143,45545,1731294928236] 2024-11-11T03:15:32,298 INFO [RS:1;59e6b60f6143:45545 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-11T03:15:32,300 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/59e6b60f6143,45545,1731294928236 already deleted, retry=false 2024-11-11T03:15:32,300 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 59e6b60f6143,45545,1731294928236 expired; onlineServers=2 2024-11-11T03:15:32,300 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [59e6b60f6143,41395,1731294928286] 2024-11-11T03:15:32,303 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/59e6b60f6143,41395,1731294928286 already deleted, retry=false 2024-11-11T03:15:32,303 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 59e6b60f6143,41395,1731294928286 expired; onlineServers=1 2024-11-11T03:15:32,306 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/default/TestHBaseWalOnEC/fb960b26c6eb2ef1ad22da807871cd25/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-11-11T03:15:32,308 INFO [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:32,309 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for fb960b26c6eb2ef1ad22da807871cd25: Waiting for close lock at 1731294932274Running coprocessor pre-close hooks at 1731294932275 (+1 ms)Disabling compacts and flushes for region at 1731294932275Disabling writes for close at 1731294932275Writing region close event to WAL at 1731294932289 (+14 ms)Running coprocessor post-close hooks at 1731294932307 (+18 ms)Closed at 1731294932308 (+1 ms) 2024-11-11T03:15:32,309 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25. 2024-11-11T03:15:32,316 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/info/2dc98713870245e99625874a31e898de is 153, key is TestHBaseWalOnEC,,1731294930421.fb960b26c6eb2ef1ad22da807871cd25./info:regioninfo/1731294930866/Put/seqid=0 2024-11-11T03:15:32,319 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,319 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,323 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1119780539_22 at /127.0.0.1:40288 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:45271:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:40288 dst: /127.0.0.1:45271 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:32,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775632_1027 (size=6637) 2024-11-11T03:15:32,328 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:32,328 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/info/2dc98713870245e99625874a31e898de 2024-11-11T03:15:32,350 INFO [regionserver/59e6b60f6143:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:32,350 INFO [regionserver/59e6b60f6143:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:32,351 INFO [regionserver/59e6b60f6143:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:32,357 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/ns/71e0acc8b16c4d618cde0d74feb1eda2 is 43, key is default/ns:d/1731294930200/Put/seqid=0 2024-11-11T03:15:32,359 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,359 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,362 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1119780539_22 at /127.0.0.1:40308 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:45271:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:40308 dst: /127.0.0.1:45271 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:32,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775616_1029 (size=5153) 2024-11-11T03:15:32,367 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:32,367 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/ns/71e0acc8b16c4d618cde0d74feb1eda2 2024-11-11T03:15:32,395 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/table/fbfd513fe35c4730834bf9024ea73fdf is 52, key is TestHBaseWalOnEC/table:state/1731294930884/Put/seqid=0 2024-11-11T03:15:32,398 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,398 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:32,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41395-0x101050601d00003, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:32,400 INFO [RS:2;59e6b60f6143:41395 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-11T03:15:32,400 INFO [RS:2;59e6b60f6143:41395 {}] regionserver.HRegionServer(1031): Exiting; stopping=59e6b60f6143,41395,1731294928286; zookeeper connection closed. 2024-11-11T03:15:32,400 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:32,400 INFO [RS:1;59e6b60f6143:45545 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-11T03:15:32,401 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45545-0x101050601d00002, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:32,401 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@26040ec3 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@26040ec3 2024-11-11T03:15:32,401 INFO [RS:1;59e6b60f6143:45545 {}] regionserver.HRegionServer(1031): Exiting; stopping=59e6b60f6143,45545,1731294928236; zookeeper connection closed. 2024-11-11T03:15:32,401 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7c0d7897 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7c0d7897 2024-11-11T03:15:32,402 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1119780539_22 at /127.0.0.1:40328 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:45271:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:40328 dst: /127.0.0.1:45271 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:32,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775600_1031 (size=5249) 2024-11-11T03:15:32,408 WARN [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:32,408 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/table/fbfd513fe35c4730834bf9024ea73fdf 2024-11-11T03:15:32,419 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/info/2dc98713870245e99625874a31e898de as hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/info/2dc98713870245e99625874a31e898de 2024-11-11T03:15:32,429 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/info/2dc98713870245e99625874a31e898de, entries=10, sequenceid=11, filesize=6.5 K 2024-11-11T03:15:32,431 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/ns/71e0acc8b16c4d618cde0d74feb1eda2 as hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/ns/71e0acc8b16c4d618cde0d74feb1eda2 2024-11-11T03:15:32,442 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/ns/71e0acc8b16c4d618cde0d74feb1eda2, entries=2, sequenceid=11, filesize=5.0 K 2024-11-11T03:15:32,444 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/.tmp/table/fbfd513fe35c4730834bf9024ea73fdf as hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/table/fbfd513fe35c4730834bf9024ea73fdf 2024-11-11T03:15:32,448 INFO [regionserver/59e6b60f6143:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-11-11T03:15:32,448 INFO [regionserver/59e6b60f6143:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-11-11T03:15:32,455 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/table/fbfd513fe35c4730834bf9024ea73fdf, entries=2, sequenceid=11, filesize=5.1 K 2024-11-11T03:15:32,457 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 180ms, sequenceid=11, compaction requested=false 2024-11-11T03:15:32,457 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-11-11T03:15:32,465 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-11-11T03:15:32,466 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-11T03:15:32,467 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-11T03:15:32,467 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731294932276Running coprocessor pre-close hooks at 1731294932276Disabling compacts and flushes for region at 1731294932276Disabling writes for close at 1731294932277 (+1 ms)Obtaining lock to block concurrent updates at 1731294932277Preparing flush snapshotting stores in 1588230740 at 1731294932277Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1731294932278 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1731294932281 (+3 ms)Flushing 1588230740/info: creating writer at 1731294932282 (+1 ms)Flushing 1588230740/info: appending metadata at 1731294932313 (+31 ms)Flushing 1588230740/info: closing flushed file at 1731294932313Flushing 1588230740/ns: creating writer at 1731294932338 (+25 ms)Flushing 1588230740/ns: appending metadata at 1731294932355 (+17 ms)Flushing 1588230740/ns: closing flushed file at 1731294932355Flushing 1588230740/table: creating writer at 1731294932377 (+22 ms)Flushing 1588230740/table: appending metadata at 1731294932394 (+17 ms)Flushing 1588230740/table: closing flushed file at 1731294932394Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5cd4ff9c: reopening flushed file at 1731294932417 (+23 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4ae37f1d: reopening flushed file at 1731294932429 (+12 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3c8d3208: reopening flushed file at 1731294932442 (+13 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 180ms, sequenceid=11, compaction requested=false at 1731294932457 (+15 ms)Writing region close event to WAL at 1731294932459 (+2 ms)Running coprocessor post-close hooks at 1731294932466 (+7 ms)Closed at 1731294932466 2024-11-11T03:15:32,467 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-11-11T03:15:32,477 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(976): stopping server 59e6b60f6143,36869,1731294928124; all regions closed. 2024-11-11T03:15:32,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_1073741829_1019 (size=2751) 2024-11-11T03:15:32,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_1073741829_1019 (size=2751) 2024-11-11T03:15:32,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_1073741829_1019 (size=2751) 2024-11-11T03:15:32,483 DEBUG [RS:0;59e6b60f6143:36869 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/oldWALs 2024-11-11T03:15:32,483 INFO [RS:0;59e6b60f6143:36869 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 59e6b60f6143%2C36869%2C1731294928124.meta:.meta(num 1731294930042) 2024-11-11T03:15:32,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_1073741828_1018 (size=1298) 2024-11-11T03:15:32,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_1073741828_1018 (size=1298) 2024-11-11T03:15:32,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_1073741828_1018 (size=1298) 2024-11-11T03:15:32,489 DEBUG [RS:0;59e6b60f6143:36869 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/oldWALs 2024-11-11T03:15:32,489 INFO [RS:0;59e6b60f6143:36869 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 59e6b60f6143%2C36869%2C1731294928124:(num 1731294929613) 2024-11-11T03:15:32,489 DEBUG [RS:0;59e6b60f6143:36869 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:32,489 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:32,489 INFO [RS:0;59e6b60f6143:36869 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-11T03:15:32,489 INFO [RS:0;59e6b60f6143:36869 {}] hbase.ChoreService(370): Chore service for: regionserver/59e6b60f6143:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-11T03:15:32,490 INFO [RS:0;59e6b60f6143:36869 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-11T03:15:32,490 INFO [regionserver/59e6b60f6143:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-11T03:15:32,490 INFO [RS:0;59e6b60f6143:36869 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36869 2024-11-11T03:15:32,492 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/59e6b60f6143,36869,1731294928124 2024-11-11T03:15:32,492 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-11T03:15:32,492 INFO [RS:0;59e6b60f6143:36869 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-11T03:15:32,494 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [59e6b60f6143,36869,1731294928124] 2024-11-11T03:15:32,495 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/59e6b60f6143,36869,1731294928124 already deleted, retry=false 2024-11-11T03:15:32,495 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 59e6b60f6143,36869,1731294928124 expired; onlineServers=0 2024-11-11T03:15:32,496 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '59e6b60f6143,42221,1731294927435' ***** 2024-11-11T03:15:32,496 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-11-11T03:15:32,496 INFO [M:0;59e6b60f6143:42221 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-11T03:15:32,496 INFO [M:0;59e6b60f6143:42221 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-11T03:15:32,496 DEBUG [M:0;59e6b60f6143:42221 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-11-11T03:15:32,496 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-11-11T03:15:32,496 DEBUG [M:0;59e6b60f6143:42221 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-11-11T03:15:32,496 DEBUG [master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.large.0-1731294929249 {}] cleaner.HFileCleaner(306): Exit Thread[master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.large.0-1731294929249,5,FailOnTimeoutGroup] 2024-11-11T03:15:32,496 DEBUG [master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.small.0-1731294929257 {}] cleaner.HFileCleaner(306): Exit Thread[master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.small.0-1731294929257,5,FailOnTimeoutGroup] 2024-11-11T03:15:32,497 INFO [M:0;59e6b60f6143:42221 {}] hbase.ChoreService(370): Chore service for: master/59e6b60f6143:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-11-11T03:15:32,497 INFO [M:0;59e6b60f6143:42221 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-11T03:15:32,497 DEBUG [M:0;59e6b60f6143:42221 {}] master.HMaster(1795): Stopping service threads 2024-11-11T03:15:32,497 INFO [M:0;59e6b60f6143:42221 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-11-11T03:15:32,497 INFO [M:0;59e6b60f6143:42221 {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-11T03:15:32,498 INFO [M:0;59e6b60f6143:42221 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-11-11T03:15:32,498 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-11-11T03:15:32,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-11-11T03:15:32,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:32,499 DEBUG [M:0;59e6b60f6143:42221 {}] zookeeper.ZKUtil(347): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-11-11T03:15:32,499 WARN [M:0;59e6b60f6143:42221 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-11-11T03:15:32,500 INFO [M:0;59e6b60f6143:42221 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/.lastflushedseqids 2024-11-11T03:15:32,508 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,509 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,511 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:40350 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:45271:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:40350 dst: /127.0.0.1:45271 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:32,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_-9223372036854775584_1033 (size=127) 2024-11-11T03:15:32,515 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:32,515 INFO [M:0;59e6b60f6143:42221 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-11-11T03:15:32,515 INFO [M:0;59e6b60f6143:42221 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-11-11T03:15:32,516 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-11T03:15:32,516 INFO [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:32,516 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:32,516 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-11T03:15:32,516 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:32,516 INFO [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.81 KB heapSize=34.10 KB 2024-11-11T03:15:32,536 DEBUG [M:0;59e6b60f6143:42221 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/bbf6c8bf8b4e4706b1cf03e18d654c6f is 82, key is hbase:meta,,1/info:regioninfo/1731294930129/Put/seqid=0 2024-11-11T03:15:32,538 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,538 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,541 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:45516 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:44249:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45516 dst: /127.0.0.1:44249 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:32,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_-9223372036854775568_1035 (size=5672) 2024-11-11T03:15:32,545 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:32,545 INFO [M:0;59e6b60f6143:42221 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/bbf6c8bf8b4e4706b1cf03e18d654c6f 2024-11-11T03:15:32,571 DEBUG [M:0;59e6b60f6143:42221 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ba00b692ebd241b0a0cff074431b0f26 is 747, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1731294930890/Put/seqid=0 2024-11-11T03:15:32,573 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,574 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,577 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:37464 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:34821:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37464 dst: /127.0.0.1:34821 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:32,581 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_-9223372036854775552_1037 (size=6437) 2024-11-11T03:15:32,582 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:32,582 INFO [M:0;59e6b60f6143:42221 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.12 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ba00b692ebd241b0a0cff074431b0f26 2024-11-11T03:15:32,594 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:32,594 INFO [RS:0;59e6b60f6143:36869 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-11T03:15:32,594 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36869-0x101050601d00001, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:32,594 INFO [RS:0;59e6b60f6143:36869 {}] regionserver.HRegionServer(1031): Exiting; stopping=59e6b60f6143,36869,1731294928124; zookeeper connection closed. 2024-11-11T03:15:32,594 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7ad2ef96 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7ad2ef96 2024-11-11T03:15:32,595 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-11-11T03:15:32,608 DEBUG [M:0;59e6b60f6143:42221 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/fb9d4302f0f2446da07123c38307c2ad is 69, key is 59e6b60f6143,36869,1731294928124/rs:state/1731294929358/Put/seqid=0 2024-11-11T03:15:32,609 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,610 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-11T03:15:32,612 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2097146950_22 at /127.0.0.1:45532 [Receiving block BP-1584767127-172.17.0.2-1731294924293:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:44249:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45532 dst: /127.0.0.1:44249 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-11T03:15:32,616 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_-9223372036854775536_1039 (size=5294) 2024-11-11T03:15:32,617 WARN [M:0;59e6b60f6143:42221 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-11T03:15:32,617 INFO [M:0;59e6b60f6143:42221 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/fb9d4302f0f2446da07123c38307c2ad 2024-11-11T03:15:32,626 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/bbf6c8bf8b4e4706b1cf03e18d654c6f as hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/bbf6c8bf8b4e4706b1cf03e18d654c6f 2024-11-11T03:15:32,635 INFO [M:0;59e6b60f6143:42221 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/bbf6c8bf8b4e4706b1cf03e18d654c6f, entries=8, sequenceid=72, filesize=5.5 K 2024-11-11T03:15:32,637 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ba00b692ebd241b0a0cff074431b0f26 as hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ba00b692ebd241b0a0cff074431b0f26 2024-11-11T03:15:32,645 INFO [M:0;59e6b60f6143:42221 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ba00b692ebd241b0a0cff074431b0f26, entries=8, sequenceid=72, filesize=6.3 K 2024-11-11T03:15:32,647 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/fb9d4302f0f2446da07123c38307c2ad as hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/fb9d4302f0f2446da07123c38307c2ad 2024-11-11T03:15:32,655 INFO [M:0;59e6b60f6143:42221 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/fb9d4302f0f2446da07123c38307c2ad, entries=3, sequenceid=72, filesize=5.2 K 2024-11-11T03:15:32,657 INFO [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.81 KB/27450, heapSize ~33.80 KB/34616, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 140ms, sequenceid=72, compaction requested=false 2024-11-11T03:15:32,658 INFO [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:32,658 DEBUG [M:0;59e6b60f6143:42221 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731294932516Disabling compacts and flushes for region at 1731294932516Disabling writes for close at 1731294932516Obtaining lock to block concurrent updates at 1731294932516Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1731294932516Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27450, getHeapSize=34856, getOffHeapSize=0, getCellsCount=85 at 1731294932517 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1731294932518 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1731294932518Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1731294932535 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1731294932535Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1731294932553 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1731294932571 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1731294932571Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1731294932590 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1731294932607 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1731294932607Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4d0b231a: reopening flushed file at 1731294932625 (+18 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2c2e4113: reopening flushed file at 1731294932635 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@587b3ed3: reopening flushed file at 1731294932645 (+10 ms)Finished flush of dataSize ~26.81 KB/27450, heapSize ~33.80 KB/34616, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 140ms, sequenceid=72, compaction requested=false at 1731294932657 (+12 ms)Writing region close event to WAL at 1731294932658 (+1 ms)Closed at 1731294932658 2024-11-11T03:15:32,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34821 is added to blk_1073741825_1011 (size=32653) 2024-11-11T03:15:32,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45271 is added to blk_1073741825_1011 (size=32653) 2024-11-11T03:15:32,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44249 is added to blk_1073741825_1011 (size=32653) 2024-11-11T03:15:32,663 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-11T03:15:32,663 INFO [M:0;59e6b60f6143:42221 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-11-11T03:15:32,663 INFO [M:0;59e6b60f6143:42221 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:42221 2024-11-11T03:15:32,663 INFO [M:0;59e6b60f6143:42221 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-11T03:15:32,766 INFO [M:0;59e6b60f6143:42221 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-11T03:15:32,766 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:32,766 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42221-0x101050601d00000, quorum=127.0.0.1:54485, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:32,770 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2e59159d{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:32,772 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@a8e922f{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-11T03:15:32,773 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-11T03:15:32,773 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@24f92c39{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-11T03:15:32,773 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@c62369b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir/,STOPPED} 2024-11-11T03:15:32,776 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-11T03:15:32,776 WARN [BP-1584767127-172.17.0.2-1731294924293 heartbeating to localhost/127.0.0.1:44063 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-11T03:15:32,776 WARN [BP-1584767127-172.17.0.2-1731294924293 heartbeating to localhost/127.0.0.1:44063 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1584767127-172.17.0.2-1731294924293 (Datanode Uuid 99217cbf-e2d6-4c92-ae21-66ce2577d1c7) service to localhost/127.0.0.1:44063 2024-11-11T03:15:32,776 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-11T03:15:32,777 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data5/current/BP-1584767127-172.17.0.2-1731294924293 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:32,777 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data6/current/BP-1584767127-172.17.0.2-1731294924293 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:32,777 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-11T03:15:32,785 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1c6b8f01{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:32,785 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11f28dd2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-11T03:15:32,785 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-11T03:15:32,786 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7fa8fa5c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-11T03:15:32,786 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6463ad04{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir/,STOPPED} 2024-11-11T03:15:32,787 WARN [BP-1584767127-172.17.0.2-1731294924293 heartbeating to localhost/127.0.0.1:44063 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-11T03:15:32,787 WARN [BP-1584767127-172.17.0.2-1731294924293 heartbeating to localhost/127.0.0.1:44063 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1584767127-172.17.0.2-1731294924293 (Datanode Uuid 0355b2ca-8a6f-4761-9367-123eaf18ca40) service to localhost/127.0.0.1:44063 2024-11-11T03:15:32,787 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-11T03:15:32,787 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-11T03:15:32,788 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data3/current/BP-1584767127-172.17.0.2-1731294924293 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:32,788 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data4/current/BP-1584767127-172.17.0.2-1731294924293 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:32,788 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-11T03:15:32,791 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4839957b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:32,791 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5306f615{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-11T03:15:32,791 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-11T03:15:32,791 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1a2478ad{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-11T03:15:32,791 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@550154bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir/,STOPPED} 2024-11-11T03:15:32,792 WARN [BP-1584767127-172.17.0.2-1731294924293 heartbeating to localhost/127.0.0.1:44063 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-11T03:15:32,793 WARN [BP-1584767127-172.17.0.2-1731294924293 heartbeating to localhost/127.0.0.1:44063 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1584767127-172.17.0.2-1731294924293 (Datanode Uuid edf79a88-1426-48ec-8141-065e335b0491) service to localhost/127.0.0.1:44063 2024-11-11T03:15:32,793 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-11T03:15:32,793 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-11T03:15:32,793 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data1/current/BP-1584767127-172.17.0.2-1731294924293 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:32,793 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/cluster_c1b01707-333f-aafd-b8e5-c875063a10c4/data/data2/current/BP-1584767127-172.17.0.2-1731294924293 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:32,794 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-11T03:15:32,805 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@76e4c45c{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-11T03:15:32,805 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-11T03:15:32,806 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-11T03:15:32,806 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-11T03:15:32,806 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir/,STOPPED} 2024-11-11T03:15:32,818 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-11-11T03:15:32,853 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-11-11T03:15:32,861 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=89 (was 159), OpenFileDescriptor=441 (was 391) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=47 (was 17) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=8029 (was 8329) 2024-11-11T03:15:32,869 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=89, OpenFileDescriptor=441, MaxFileDescriptor=1048576, SystemLoadAverage=47, ProcessCount=11, AvailableMemoryMB=8029 2024-11-11T03:15:32,869 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-11-11T03:15:32,869 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.log.dir so I do NOT create it in target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484 2024-11-11T03:15:32,869 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/ce1ed874-c66b-5796-92ba-d2b45d50717e/hadoop.tmp.dir so I do NOT create it in target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484 2024-11-11T03:15:32,870 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4, deleteOnExit=true 2024-11-11T03:15:32,870 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-11-11T03:15:32,870 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/test.cache.data in system properties and HBase conf 2024-11-11T03:15:32,870 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.tmp.dir in system properties and HBase conf 2024-11-11T03:15:32,870 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir in system properties and HBase conf 2024-11-11T03:15:32,870 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/mapreduce.cluster.local.dir in system properties and HBase conf 2024-11-11T03:15:32,871 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-11-11T03:15:32,871 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-11-11T03:15:32,871 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-11-11T03:15:32,871 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-11-11T03:15:32,871 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-11-11T03:15:32,872 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-11-11T03:15:32,872 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-11T03:15:32,872 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-11-11T03:15:32,872 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-11-11T03:15:32,872 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-11T03:15:32,872 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-11T03:15:32,872 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-11-11T03:15:32,873 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/nfs.dump.dir in system properties and HBase conf 2024-11-11T03:15:32,873 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/java.io.tmpdir in system properties and HBase conf 2024-11-11T03:15:32,873 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-11T03:15:32,873 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-11-11T03:15:32,873 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-11-11T03:15:32,982 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:32,988 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-11T03:15:32,990 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-11T03:15:32,990 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-11T03:15:32,990 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-11T03:15:32,991 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:32,992 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@49b2b984{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir/,AVAILABLE} 2024-11-11T03:15:32,992 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@715f09c8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-11T03:15:33,115 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3b25f894{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/java.io.tmpdir/jetty-localhost-34329-hadoop-hdfs-3_4_1-tests_jar-_-any-10306468111902147603/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-11T03:15:33,116 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@53797dc3{HTTP/1.1, (http/1.1)}{localhost:34329} 2024-11-11T03:15:33,116 INFO [Time-limited test {}] server.Server(415): Started @10676ms 2024-11-11T03:15:33,204 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:33,207 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-11T03:15:33,208 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-11T03:15:33,208 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-11T03:15:33,208 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-11T03:15:33,208 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5a6744cf{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir/,AVAILABLE} 2024-11-11T03:15:33,209 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7c2dd4e6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-11T03:15:33,323 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@14abb266{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/java.io.tmpdir/jetty-localhost-39537-hadoop-hdfs-3_4_1-tests_jar-_-any-2460852758270772921/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:33,324 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@348ccaab{HTTP/1.1, (http/1.1)}{localhost:39537} 2024-11-11T03:15:33,324 INFO [Time-limited test {}] server.Server(415): Started @10884ms 2024-11-11T03:15:33,325 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-11T03:15:33,359 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:33,362 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-11T03:15:33,365 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-11T03:15:33,365 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-11T03:15:33,365 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-11T03:15:33,365 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@69cb0b1f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir/,AVAILABLE} 2024-11-11T03:15:33,366 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6f8ad177{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-11T03:15:33,435 WARN [Thread-522 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data1/current/BP-1808564834-172.17.0.2-1731294932919/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:33,435 WARN [Thread-523 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data2/current/BP-1808564834-172.17.0.2-1731294932919/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:33,459 WARN [Thread-501 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-11T03:15:33,463 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd0f575d64d48f551 with lease ID 0x57b5bd8e096c87c4: Processing first storage report for DS-6d1e28f9-acd7-44ac-94c6-d6aa780a647c from datanode DatanodeRegistration(127.0.0.1:44627, datanodeUuid=e2f14b05-80b8-45e2-8c6a-1a5d8d53a212, infoPort=42701, infoSecurePort=0, ipcPort=44339, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919) 2024-11-11T03:15:33,463 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd0f575d64d48f551 with lease ID 0x57b5bd8e096c87c4: from storage DS-6d1e28f9-acd7-44ac-94c6-d6aa780a647c node DatanodeRegistration(127.0.0.1:44627, datanodeUuid=e2f14b05-80b8-45e2-8c6a-1a5d8d53a212, infoPort=42701, infoSecurePort=0, ipcPort=44339, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-11T03:15:33,463 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd0f575d64d48f551 with lease ID 0x57b5bd8e096c87c4: Processing first storage report for DS-ed46f5c0-b44a-4227-b7d9-df618c27f7d9 from datanode DatanodeRegistration(127.0.0.1:44627, datanodeUuid=e2f14b05-80b8-45e2-8c6a-1a5d8d53a212, infoPort=42701, infoSecurePort=0, ipcPort=44339, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919) 2024-11-11T03:15:33,463 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd0f575d64d48f551 with lease ID 0x57b5bd8e096c87c4: from storage DS-ed46f5c0-b44a-4227-b7d9-df618c27f7d9 node DatanodeRegistration(127.0.0.1:44627, datanodeUuid=e2f14b05-80b8-45e2-8c6a-1a5d8d53a212, infoPort=42701, infoSecurePort=0, ipcPort=44339, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-11T03:15:33,499 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@609afea4{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/java.io.tmpdir/jetty-localhost-35983-hadoop-hdfs-3_4_1-tests_jar-_-any-11725302900140593019/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:33,499 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@527d4f04{HTTP/1.1, (http/1.1)}{localhost:35983} 2024-11-11T03:15:33,499 INFO [Time-limited test {}] server.Server(415): Started @11059ms 2024-11-11T03:15:33,501 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-11T03:15:33,532 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-11T03:15:33,535 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-11T03:15:33,536 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-11T03:15:33,536 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-11T03:15:33,536 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-11T03:15:33,537 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7adc0795{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir/,AVAILABLE} 2024-11-11T03:15:33,537 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@739551bc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-11T03:15:33,600 WARN [Thread-557 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data3/current/BP-1808564834-172.17.0.2-1731294932919/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:33,600 WARN [Thread-558 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data4/current/BP-1808564834-172.17.0.2-1731294932919/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:33,617 WARN [Thread-537 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-11T03:15:33,621 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3e336b1b7a314ef6 with lease ID 0x57b5bd8e096c87c5: Processing first storage report for DS-8b570462-f1cb-4908-805b-511f391cdb98 from datanode DatanodeRegistration(127.0.0.1:41935, datanodeUuid=d7bfa18f-a190-478f-a4dc-3c02c56ba7f4, infoPort=37869, infoSecurePort=0, ipcPort=36033, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919) 2024-11-11T03:15:33,621 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3e336b1b7a314ef6 with lease ID 0x57b5bd8e096c87c5: from storage DS-8b570462-f1cb-4908-805b-511f391cdb98 node DatanodeRegistration(127.0.0.1:41935, datanodeUuid=d7bfa18f-a190-478f-a4dc-3c02c56ba7f4, infoPort=37869, infoSecurePort=0, ipcPort=36033, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-11T03:15:33,621 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3e336b1b7a314ef6 with lease ID 0x57b5bd8e096c87c5: Processing first storage report for DS-f2accb31-114a-4613-b5a6-a60649e7213e from datanode DatanodeRegistration(127.0.0.1:41935, datanodeUuid=d7bfa18f-a190-478f-a4dc-3c02c56ba7f4, infoPort=37869, infoSecurePort=0, ipcPort=36033, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919) 2024-11-11T03:15:33,621 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3e336b1b7a314ef6 with lease ID 0x57b5bd8e096c87c5: from storage DS-f2accb31-114a-4613-b5a6-a60649e7213e node DatanodeRegistration(127.0.0.1:41935, datanodeUuid=d7bfa18f-a190-478f-a4dc-3c02c56ba7f4, infoPort=37869, infoSecurePort=0, ipcPort=36033, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-11T03:15:33,656 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@25c162fe{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/java.io.tmpdir/jetty-localhost-38715-hadoop-hdfs-3_4_1-tests_jar-_-any-1236365998481153171/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:33,656 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1da1746e{HTTP/1.1, (http/1.1)}{localhost:38715} 2024-11-11T03:15:33,656 INFO [Time-limited test {}] server.Server(415): Started @11217ms 2024-11-11T03:15:33,658 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-11T03:15:33,764 WARN [Thread-584 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data6/current/BP-1808564834-172.17.0.2-1731294932919/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:33,764 WARN [Thread-583 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data5/current/BP-1808564834-172.17.0.2-1731294932919/current, will proceed with Du for space computation calculation, 2024-11-11T03:15:33,786 WARN [Thread-572 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-11T03:15:33,789 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xa1527d31d83d9eff with lease ID 0x57b5bd8e096c87c6: Processing first storage report for DS-c1fdb2ea-97ad-4e32-a46f-53a13545ae5c from datanode DatanodeRegistration(127.0.0.1:41783, datanodeUuid=b1a66a44-013b-4763-aedf-4dd4bc268fa6, infoPort=38055, infoSecurePort=0, ipcPort=45223, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919) 2024-11-11T03:15:33,789 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa1527d31d83d9eff with lease ID 0x57b5bd8e096c87c6: from storage DS-c1fdb2ea-97ad-4e32-a46f-53a13545ae5c node DatanodeRegistration(127.0.0.1:41783, datanodeUuid=b1a66a44-013b-4763-aedf-4dd4bc268fa6, infoPort=38055, infoSecurePort=0, ipcPort=45223, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-11T03:15:33,790 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xa1527d31d83d9eff with lease ID 0x57b5bd8e096c87c6: Processing first storage report for DS-21234118-8c6e-4ea0-b4c5-d3e91e9fabd5 from datanode DatanodeRegistration(127.0.0.1:41783, datanodeUuid=b1a66a44-013b-4763-aedf-4dd4bc268fa6, infoPort=38055, infoSecurePort=0, ipcPort=45223, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919) 2024-11-11T03:15:33,790 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa1527d31d83d9eff with lease ID 0x57b5bd8e096c87c6: from storage DS-21234118-8c6e-4ea0-b4c5-d3e91e9fabd5 node DatanodeRegistration(127.0.0.1:41783, datanodeUuid=b1a66a44-013b-4763-aedf-4dd4bc268fa6, infoPort=38055, infoSecurePort=0, ipcPort=45223, storageInfo=lv=-57;cid=testClusterID;nsid=849370717;c=1731294932919), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-11T03:15:33,885 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484 2024-11-11T03:15:33,888 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/zookeeper_0, clientPort=60206, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-11-11T03:15:33,889 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=60206 2024-11-11T03:15:33,889 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:33,891 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:33,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741825_1001 (size=7) 2024-11-11T03:15:33,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741825_1001 (size=7) 2024-11-11T03:15:33,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741825_1001 (size=7) 2024-11-11T03:15:33,906 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605 with version=8 2024-11-11T03:15:33,906 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:44063/user/jenkins/test-data/e1d7c0ff-1ef4-2c68-118b-8df462507939/hbase-staging 2024-11-11T03:15:33,908 INFO [Time-limited test {}] client.ConnectionUtils(128): master/59e6b60f6143:0 server-side Connection retries=45 2024-11-11T03:15:33,908 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,908 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,908 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-11T03:15:33,908 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,908 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-11T03:15:33,909 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-11-11T03:15:33,909 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-11T03:15:33,909 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41405 2024-11-11T03:15:33,911 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:41405 connecting to ZooKeeper ensemble=127.0.0.1:60206 2024-11-11T03:15:33,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:414050x0, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-11T03:15:33,919 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:41405-0x10105061e0f0000 connected 2024-11-11T03:15:33,935 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:33,937 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:33,939 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:33,939 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605, hbase.cluster.distributed=false 2024-11-11T03:15:33,941 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-11T03:15:33,941 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41405 2024-11-11T03:15:33,942 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41405 2024-11-11T03:15:33,942 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41405 2024-11-11T03:15:33,944 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41405 2024-11-11T03:15:33,944 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41405 2024-11-11T03:15:33,960 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/59e6b60f6143:0 server-side Connection retries=45 2024-11-11T03:15:33,960 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,960 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,960 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-11T03:15:33,960 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,960 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-11T03:15:33,960 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-11T03:15:33,960 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-11T03:15:33,961 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41097 2024-11-11T03:15:33,962 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:41097 connecting to ZooKeeper ensemble=127.0.0.1:60206 2024-11-11T03:15:33,963 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:33,965 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:33,969 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:410970x0, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-11T03:15:33,969 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:41097-0x10105061e0f0001 connected 2024-11-11T03:15:33,969 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:33,970 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-11T03:15:33,970 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-11T03:15:33,971 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-11T03:15:33,972 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-11T03:15:33,973 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41097 2024-11-11T03:15:33,973 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41097 2024-11-11T03:15:33,973 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41097 2024-11-11T03:15:33,973 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41097 2024-11-11T03:15:33,974 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41097 2024-11-11T03:15:33,989 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/59e6b60f6143:0 server-side Connection retries=45 2024-11-11T03:15:33,989 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,989 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,989 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-11T03:15:33,989 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:33,989 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-11T03:15:33,989 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-11T03:15:33,989 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-11T03:15:33,990 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41049 2024-11-11T03:15:33,991 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:41049 connecting to ZooKeeper ensemble=127.0.0.1:60206 2024-11-11T03:15:33,992 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:33,993 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:33,997 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:410490x0, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-11T03:15:33,998 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:410490x0, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:33,998 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:41049-0x10105061e0f0002 connected 2024-11-11T03:15:33,998 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-11T03:15:33,999 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-11T03:15:33,999 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-11T03:15:34,000 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-11T03:15:34,001 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41049 2024-11-11T03:15:34,001 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41049 2024-11-11T03:15:34,001 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41049 2024-11-11T03:15:34,002 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41049 2024-11-11T03:15:34,002 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41049 2024-11-11T03:15:34,017 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/59e6b60f6143:0 server-side Connection retries=45 2024-11-11T03:15:34,018 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:34,018 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:34,018 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-11T03:15:34,018 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-11T03:15:34,018 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-11T03:15:34,018 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-11T03:15:34,018 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-11T03:15:34,019 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:35667 2024-11-11T03:15:34,020 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:35667 connecting to ZooKeeper ensemble=127.0.0.1:60206 2024-11-11T03:15:34,021 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:34,022 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:34,031 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:356670x0, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-11T03:15:34,032 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:356670x0, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:34,032 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35667-0x10105061e0f0003 connected 2024-11-11T03:15:34,032 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-11T03:15:34,033 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-11T03:15:34,033 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-11T03:15:34,035 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-11T03:15:34,035 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35667 2024-11-11T03:15:34,035 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35667 2024-11-11T03:15:34,036 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35667 2024-11-11T03:15:34,036 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35667 2024-11-11T03:15:34,036 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35667 2024-11-11T03:15:34,049 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;59e6b60f6143:41405 2024-11-11T03:15:34,049 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/59e6b60f6143,41405,1731294933908 2024-11-11T03:15:34,051 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:34,051 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:34,051 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:34,051 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:34,052 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/59e6b60f6143,41405,1731294933908 2024-11-11T03:15:34,054 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-11T03:15:34,054 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-11T03:15:34,054 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,054 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,054 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,055 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-11T03:15:34,055 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,057 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-11-11T03:15:34,057 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/59e6b60f6143,41405,1731294933908 from backup master directory 2024-11-11T03:15:34,059 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/59e6b60f6143,41405,1731294933908 2024-11-11T03:15:34,059 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:34,059 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:34,059 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:34,059 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-11T03:15:34,059 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-11T03:15:34,059 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=59e6b60f6143,41405,1731294933908 2024-11-11T03:15:34,067 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/hbase.id] with ID: b16762e3-d680-4989-9837-cbc94cfa07e2 2024-11-11T03:15:34,067 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/.tmp/hbase.id 2024-11-11T03:15:34,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741826_1002 (size=42) 2024-11-11T03:15:34,076 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741826_1002 (size=42) 2024-11-11T03:15:34,077 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741826_1002 (size=42) 2024-11-11T03:15:34,077 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/.tmp/hbase.id]:[hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/hbase.id] 2024-11-11T03:15:34,094 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-11T03:15:34,094 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-11-11T03:15:34,096 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 1ms. 2024-11-11T03:15:34,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,113 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741827_1003 (size=196) 2024-11-11T03:15:34,114 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741827_1003 (size=196) 2024-11-11T03:15:34,115 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741827_1003 (size=196) 2024-11-11T03:15:34,115 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-11T03:15:34,116 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-11-11T03:15:34,119 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-11T03:15:34,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741828_1004 (size=1189) 2024-11-11T03:15:34,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741828_1004 (size=1189) 2024-11-11T03:15:34,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741828_1004 (size=1189) 2024-11-11T03:15:34,135 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store 2024-11-11T03:15:34,148 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741829_1005 (size=34) 2024-11-11T03:15:34,148 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741829_1005 (size=34) 2024-11-11T03:15:34,148 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741829_1005 (size=34) 2024-11-11T03:15:34,149 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:34,149 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-11T03:15:34,149 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:34,149 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:34,149 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-11T03:15:34,149 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:34,149 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:34,149 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731294934149Disabling compacts and flushes for region at 1731294934149Disabling writes for close at 1731294934149Writing region close event to WAL at 1731294934149Closed at 1731294934149 2024-11-11T03:15:34,151 WARN [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/.initializing 2024-11-11T03:15:34,151 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/WALs/59e6b60f6143,41405,1731294933908 2024-11-11T03:15:34,155 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C41405%2C1731294933908, suffix=, logDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/WALs/59e6b60f6143,41405,1731294933908, archiveDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/oldWALs, maxLogs=10 2024-11-11T03:15:34,156 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 59e6b60f6143%2C41405%2C1731294933908.1731294934155 2024-11-11T03:15:34,166 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/WALs/59e6b60f6143,41405,1731294933908/59e6b60f6143%2C41405%2C1731294933908.1731294934155 2024-11-11T03:15:34,169 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37869:37869),(127.0.0.1/127.0.0.1:42701:42701),(127.0.0.1/127.0.0.1:38055:38055)] 2024-11-11T03:15:34,170 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-11-11T03:15:34,170 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:34,171 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,171 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,173 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,175 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-11-11T03:15:34,175 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,175 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,175 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,177 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-11-11T03:15:34,177 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,178 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-11T03:15:34,178 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,180 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-11-11T03:15:34,180 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,181 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-11T03:15:34,181 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,182 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-11-11T03:15:34,182 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,183 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-11T03:15:34,183 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,184 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,184 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,186 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,186 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,187 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-11T03:15:34,188 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-11-11T03:15:34,191 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-11T03:15:34,191 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=62291187, jitterRate=-0.0717889815568924}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-11T03:15:34,192 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1731294934171Initializing all the Stores at 1731294934172 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294934172Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294934173 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294934173Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294934173Cleaning up temporary data from old regions at 1731294934186 (+13 ms)Region opened successfully at 1731294934192 (+6 ms) 2024-11-11T03:15:34,193 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-11-11T03:15:34,197 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@31d9c3bd, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=59e6b60f6143/172.17.0.2:0 2024-11-11T03:15:34,198 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-11-11T03:15:34,199 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-11-11T03:15:34,199 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-11-11T03:15:34,199 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-11-11T03:15:34,199 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-11-11T03:15:34,200 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-11-11T03:15:34,200 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-11-11T03:15:34,202 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-11-11T03:15:34,203 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-11-11T03:15:34,205 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-11-11T03:15:34,205 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-11-11T03:15:34,206 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-11-11T03:15:34,208 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-11-11T03:15:34,208 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-11-11T03:15:34,209 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-11-11T03:15:34,210 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-11-11T03:15:34,211 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-11-11T03:15:34,213 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-11-11T03:15:34,215 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-11-11T03:15:34,217 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-11-11T03:15:34,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:34,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:34,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:34,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:34,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,220 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=59e6b60f6143,41405,1731294933908, sessionid=0x10105061e0f0000, setting cluster-up flag (Was=false) 2024-11-11T03:15:34,224 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,224 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,224 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,224 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,229 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-11-11T03:15:34,230 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=59e6b60f6143,41405,1731294933908 2024-11-11T03:15:34,235 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,235 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,235 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,235 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,241 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-11-11T03:15:34,242 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=59e6b60f6143,41405,1731294933908 2024-11-11T03:15:34,243 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-11-11T03:15:34,246 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-11-11T03:15:34,246 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-11-11T03:15:34,246 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-11-11T03:15:34,246 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 59e6b60f6143,41405,1731294933908 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-11-11T03:15:34,248 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/59e6b60f6143:0, corePoolSize=5, maxPoolSize=5 2024-11-11T03:15:34,248 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/59e6b60f6143:0, corePoolSize=5, maxPoolSize=5 2024-11-11T03:15:34,248 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/59e6b60f6143:0, corePoolSize=5, maxPoolSize=5 2024-11-11T03:15:34,248 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/59e6b60f6143:0, corePoolSize=5, maxPoolSize=5 2024-11-11T03:15:34,248 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/59e6b60f6143:0, corePoolSize=10, maxPoolSize=10 2024-11-11T03:15:34,248 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,248 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/59e6b60f6143:0, corePoolSize=2, maxPoolSize=2 2024-11-11T03:15:34,248 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1731294964251 2024-11-11T03:15:34,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-11-11T03:15:34,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-11-11T03:15:34,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-11-11T03:15:34,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-11-11T03:15:34,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-11-11T03:15:34,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-11-11T03:15:34,251 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,252 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-11-11T03:15:34,252 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-11-11T03:15:34,252 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-11-11T03:15:34,252 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-11T03:15:34,252 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-11-11T03:15:34,252 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-11-11T03:15:34,252 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-11-11T03:15:34,254 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,254 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-11-11T03:15:34,256 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.large.0-1731294934252,5,FailOnTimeoutGroup] 2024-11-11T03:15:34,256 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.small.0-1731294934256,5,FailOnTimeoutGroup] 2024-11-11T03:15:34,256 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,256 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-11-11T03:15:34,256 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,257 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,265 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741831_1007 (size=1321) 2024-11-11T03:15:34,265 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741831_1007 (size=1321) 2024-11-11T03:15:34,266 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741831_1007 (size=1321) 2024-11-11T03:15:34,267 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-11-11T03:15:34,267 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605 2024-11-11T03:15:34,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741832_1008 (size=32) 2024-11-11T03:15:34,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741832_1008 (size=32) 2024-11-11T03:15:34,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741832_1008 (size=32) 2024-11-11T03:15:34,282 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:34,284 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-11T03:15:34,286 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-11T03:15:34,286 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,287 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,287 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-11T03:15:34,288 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-11T03:15:34,288 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,289 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,289 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-11T03:15:34,290 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-11T03:15:34,290 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,291 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,291 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-11T03:15:34,292 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-11T03:15:34,293 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,293 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,293 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-11T03:15:34,294 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740 2024-11-11T03:15:34,295 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740 2024-11-11T03:15:34,296 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-11T03:15:34,296 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-11T03:15:34,297 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-11T03:15:34,299 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-11T03:15:34,302 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-11T03:15:34,302 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71012426, jitterRate=0.058167606592178345}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-11T03:15:34,303 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1731294934283Initializing all the Stores at 1731294934284 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294934284Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294934284Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294934284Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294934284Cleaning up temporary data from old regions at 1731294934296 (+12 ms)Region opened successfully at 1731294934303 (+7 ms) 2024-11-11T03:15:34,303 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-11T03:15:34,304 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-11T03:15:34,304 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-11T03:15:34,304 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-11T03:15:34,304 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-11T03:15:34,304 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-11T03:15:34,304 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731294934303Disabling compacts and flushes for region at 1731294934303Disabling writes for close at 1731294934304 (+1 ms)Writing region close event to WAL at 1731294934304Closed at 1731294934304 2024-11-11T03:15:34,306 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-11T03:15:34,306 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-11-11T03:15:34,307 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-11-11T03:15:34,309 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-11T03:15:34,310 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-11-11T03:15:34,338 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(746): ClusterId : b16762e3-d680-4989-9837-cbc94cfa07e2 2024-11-11T03:15:34,338 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(746): ClusterId : b16762e3-d680-4989-9837-cbc94cfa07e2 2024-11-11T03:15:34,338 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(746): ClusterId : b16762e3-d680-4989-9837-cbc94cfa07e2 2024-11-11T03:15:34,338 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-11T03:15:34,338 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-11T03:15:34,338 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-11T03:15:34,341 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-11T03:15:34,341 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-11T03:15:34,341 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-11T03:15:34,341 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-11T03:15:34,341 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-11T03:15:34,341 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-11T03:15:34,344 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-11T03:15:34,344 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-11T03:15:34,344 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-11T03:15:34,344 DEBUG [RS:1;59e6b60f6143:41049 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@27abcaca, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=59e6b60f6143/172.17.0.2:0 2024-11-11T03:15:34,345 DEBUG [RS:0;59e6b60f6143:41097 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3cbedcda, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=59e6b60f6143/172.17.0.2:0 2024-11-11T03:15:34,345 DEBUG [RS:2;59e6b60f6143:35667 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2bdbde2c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=59e6b60f6143/172.17.0.2:0 2024-11-11T03:15:34,357 DEBUG [RS:2;59e6b60f6143:35667 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;59e6b60f6143:35667 2024-11-11T03:15:34,357 DEBUG [RS:0;59e6b60f6143:41097 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;59e6b60f6143:41097 2024-11-11T03:15:34,357 DEBUG [RS:1;59e6b60f6143:41049 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;59e6b60f6143:41049 2024-11-11T03:15:34,357 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-11T03:15:34,357 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-11T03:15:34,357 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-11T03:15:34,357 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-11T03:15:34,357 DEBUG [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-11T03:15:34,357 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-11T03:15:34,357 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-11T03:15:34,357 DEBUG [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-11T03:15:34,357 DEBUG [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-11T03:15:34,358 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(2659): reportForDuty to master=59e6b60f6143,41405,1731294933908 with port=41097, startcode=1731294933960 2024-11-11T03:15:34,358 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(2659): reportForDuty to master=59e6b60f6143,41405,1731294933908 with port=35667, startcode=1731294934017 2024-11-11T03:15:34,358 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(2659): reportForDuty to master=59e6b60f6143,41405,1731294933908 with port=41049, startcode=1731294933989 2024-11-11T03:15:34,359 DEBUG [RS:2;59e6b60f6143:35667 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-11T03:15:34,359 DEBUG [RS:0;59e6b60f6143:41097 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-11T03:15:34,359 DEBUG [RS:1;59e6b60f6143:41049 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-11T03:15:34,361 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37585, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-11-11T03:15:34,361 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55327, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-11-11T03:15:34,361 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:39591, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-11-11T03:15:34,362 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41405 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,362 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41405 {}] master.ServerManager(517): Registering regionserver=59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,364 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41405 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 59e6b60f6143,35667,1731294934017 2024-11-11T03:15:34,364 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41405 {}] master.ServerManager(517): Registering regionserver=59e6b60f6143,35667,1731294934017 2024-11-11T03:15:34,364 DEBUG [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605 2024-11-11T03:15:34,364 DEBUG [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:38147 2024-11-11T03:15:34,364 DEBUG [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-11T03:15:34,366 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41405 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 59e6b60f6143,41097,1731294933960 2024-11-11T03:15:34,366 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41405 {}] master.ServerManager(517): Registering regionserver=59e6b60f6143,41097,1731294933960 2024-11-11T03:15:34,366 DEBUG [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605 2024-11-11T03:15:34,366 DEBUG [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:38147 2024-11-11T03:15:34,366 DEBUG [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-11T03:15:34,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-11T03:15:34,368 DEBUG [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605 2024-11-11T03:15:34,368 DEBUG [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:38147 2024-11-11T03:15:34,368 DEBUG [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-11T03:15:34,371 DEBUG [RS:1;59e6b60f6143:41049 {}] zookeeper.ZKUtil(111): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,371 WARN [RS:1;59e6b60f6143:41049 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-11T03:15:34,372 INFO [RS:1;59e6b60f6143:41049 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-11T03:15:34,372 DEBUG [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,373 DEBUG [RS:2;59e6b60f6143:35667 {}] zookeeper.ZKUtil(111): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/59e6b60f6143,35667,1731294934017 2024-11-11T03:15:34,373 WARN [RS:2;59e6b60f6143:35667 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-11T03:15:34,373 INFO [RS:2;59e6b60f6143:35667 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-11T03:15:34,373 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [59e6b60f6143,41049,1731294933989] 2024-11-11T03:15:34,373 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [59e6b60f6143,35667,1731294934017] 2024-11-11T03:15:34,373 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [59e6b60f6143,41097,1731294933960] 2024-11-11T03:15:34,373 DEBUG [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,35667,1731294934017 2024-11-11T03:15:34,373 DEBUG [RS:0;59e6b60f6143:41097 {}] zookeeper.ZKUtil(111): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/59e6b60f6143,41097,1731294933960 2024-11-11T03:15:34,373 WARN [RS:0;59e6b60f6143:41097 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-11T03:15:34,373 INFO [RS:0;59e6b60f6143:41097 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-11T03:15:34,373 DEBUG [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,41097,1731294933960 2024-11-11T03:15:34,376 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-11T03:15:34,378 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-11T03:15:34,379 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-11T03:15:34,379 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-11T03:15:34,379 INFO [RS:1;59e6b60f6143:41049 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-11T03:15:34,379 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,380 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-11T03:15:34,384 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-11T03:15:34,384 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,384 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-11T03:15:34,384 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/59e6b60f6143:0, corePoolSize=2, maxPoolSize=2 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:34,385 DEBUG [RS:1;59e6b60f6143:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:34,387 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-11T03:15:34,388 INFO [RS:2;59e6b60f6143:35667 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-11T03:15:34,388 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,390 INFO [RS:0;59e6b60f6143:41097 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-11T03:15:34,390 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-11T03:15:34,390 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,390 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-11T03:15:34,391 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,391 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,391 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,391 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,391 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,391 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,391 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41049,1731294933989-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/59e6b60f6143:0, corePoolSize=2, maxPoolSize=2 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,391 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,391 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-11T03:15:34,392 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,392 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,392 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,392 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:34,392 DEBUG [RS:2;59e6b60f6143:35667 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:34,392 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-11T03:15:34,392 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/59e6b60f6143:0, corePoolSize=2, maxPoolSize=2 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/59e6b60f6143:0, corePoolSize=1, maxPoolSize=1 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:34,393 DEBUG [RS:0;59e6b60f6143:41097 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0, corePoolSize=3, maxPoolSize=3 2024-11-11T03:15:34,396 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,35667,1731294934017-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,397 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41097,1731294933960-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-11T03:15:34,409 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-11T03:15:34,409 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41049,1731294933989-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,409 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,409 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.Replication(171): 59e6b60f6143,41049,1731294933989 started 2024-11-11T03:15:34,412 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-11T03:15:34,412 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,35667,1731294934017-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,412 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-11T03:15:34,412 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,412 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41097,1731294933960-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,412 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.Replication(171): 59e6b60f6143,35667,1731294934017 started 2024-11-11T03:15:34,412 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,413 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.Replication(171): 59e6b60f6143,41097,1731294933960 started 2024-11-11T03:15:34,425 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,425 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1482): Serving as 59e6b60f6143,41049,1731294933989, RpcServer on 59e6b60f6143/172.17.0.2:41049, sessionid=0x10105061e0f0002 2024-11-11T03:15:34,425 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-11T03:15:34,425 DEBUG [RS:1;59e6b60f6143:41049 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,425 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,41049,1731294933989' 2024-11-11T03:15:34,425 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-11T03:15:34,426 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-11T03:15:34,426 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-11T03:15:34,426 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-11T03:15:34,426 DEBUG [RS:1;59e6b60f6143:41049 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,426 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,41049,1731294933989' 2024-11-11T03:15:34,426 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-11T03:15:34,427 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-11T03:15:34,427 DEBUG [RS:1;59e6b60f6143:41049 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-11T03:15:34,427 INFO [RS:1;59e6b60f6143:41049 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-11T03:15:34,427 INFO [RS:1;59e6b60f6143:41049 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-11T03:15:34,427 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,428 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(1482): Serving as 59e6b60f6143,35667,1731294934017, RpcServer on 59e6b60f6143/172.17.0.2:35667, sessionid=0x10105061e0f0003 2024-11-11T03:15:34,428 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-11T03:15:34,428 DEBUG [RS:2;59e6b60f6143:35667 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 59e6b60f6143,35667,1731294934017 2024-11-11T03:15:34,428 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,428 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,35667,1731294934017' 2024-11-11T03:15:34,428 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-11T03:15:34,428 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1482): Serving as 59e6b60f6143,41097,1731294933960, RpcServer on 59e6b60f6143/172.17.0.2:41097, sessionid=0x10105061e0f0001 2024-11-11T03:15:34,428 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-11T03:15:34,428 DEBUG [RS:0;59e6b60f6143:41097 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 59e6b60f6143,41097,1731294933960 2024-11-11T03:15:34,428 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,41097,1731294933960' 2024-11-11T03:15:34,428 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-11T03:15:34,428 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-11T03:15:34,429 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-11T03:15:34,429 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-11T03:15:34,429 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-11T03:15:34,429 DEBUG [RS:2;59e6b60f6143:35667 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 59e6b60f6143,35667,1731294934017 2024-11-11T03:15:34,429 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,35667,1731294934017' 2024-11-11T03:15:34,429 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-11T03:15:34,429 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-11T03:15:34,429 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-11T03:15:34,429 DEBUG [RS:0;59e6b60f6143:41097 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 59e6b60f6143,41097,1731294933960 2024-11-11T03:15:34,429 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '59e6b60f6143,41097,1731294933960' 2024-11-11T03:15:34,429 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-11T03:15:34,429 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-11T03:15:34,430 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-11T03:15:34,430 DEBUG [RS:2;59e6b60f6143:35667 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-11T03:15:34,430 INFO [RS:2;59e6b60f6143:35667 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-11T03:15:34,430 INFO [RS:2;59e6b60f6143:35667 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-11T03:15:34,430 DEBUG [RS:0;59e6b60f6143:41097 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-11T03:15:34,430 INFO [RS:0;59e6b60f6143:41097 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-11T03:15:34,430 INFO [RS:0;59e6b60f6143:41097 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-11T03:15:34,460 WARN [59e6b60f6143:41405 {}] assignment.AssignmentManager(2443): No servers available; cannot place 1 unassigned regions. 2024-11-11T03:15:34,530 INFO [RS:1;59e6b60f6143:41049 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C41049%2C1731294933989, suffix=, logDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,41049,1731294933989, archiveDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/oldWALs, maxLogs=32 2024-11-11T03:15:34,532 INFO [RS:2;59e6b60f6143:35667 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C35667%2C1731294934017, suffix=, logDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,35667,1731294934017, archiveDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/oldWALs, maxLogs=32 2024-11-11T03:15:34,532 INFO [RS:1;59e6b60f6143:41049 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 59e6b60f6143%2C41049%2C1731294933989.1731294934532 2024-11-11T03:15:34,532 INFO [RS:0;59e6b60f6143:41097 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C41097%2C1731294933960, suffix=, logDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,41097,1731294933960, archiveDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/oldWALs, maxLogs=32 2024-11-11T03:15:34,537 INFO [RS:0;59e6b60f6143:41097 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 59e6b60f6143%2C41097%2C1731294933960.1731294934536 2024-11-11T03:15:34,537 INFO [RS:2;59e6b60f6143:35667 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 59e6b60f6143%2C35667%2C1731294934017.1731294934537 2024-11-11T03:15:34,542 INFO [RS:1;59e6b60f6143:41049 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,41049,1731294933989/59e6b60f6143%2C41049%2C1731294933989.1731294934532 2024-11-11T03:15:34,548 DEBUG [RS:1;59e6b60f6143:41049 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38055:38055),(127.0.0.1/127.0.0.1:42701:42701),(127.0.0.1/127.0.0.1:37869:37869)] 2024-11-11T03:15:34,548 INFO [RS:0;59e6b60f6143:41097 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,41097,1731294933960/59e6b60f6143%2C41097%2C1731294933960.1731294934536 2024-11-11T03:15:34,548 INFO [RS:2;59e6b60f6143:35667 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,35667,1731294934017/59e6b60f6143%2C35667%2C1731294934017.1731294934537 2024-11-11T03:15:34,550 DEBUG [RS:0;59e6b60f6143:41097 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:42701:42701),(127.0.0.1/127.0.0.1:37869:37869),(127.0.0.1/127.0.0.1:38055:38055)] 2024-11-11T03:15:34,550 DEBUG [RS:2;59e6b60f6143:35667 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38055:38055),(127.0.0.1/127.0.0.1:42701:42701),(127.0.0.1/127.0.0.1:37869:37869)] 2024-11-11T03:15:34,711 DEBUG [59e6b60f6143:41405 {}] assignment.AssignmentManager(2464): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-11-11T03:15:34,711 DEBUG [59e6b60f6143:41405 {}] balancer.BalancerClusterState(204): Hosts are {59e6b60f6143=0} racks are {/default-rack=0} 2024-11-11T03:15:34,713 DEBUG [59e6b60f6143:41405 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-11T03:15:34,713 DEBUG [59e6b60f6143:41405 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-11T03:15:34,713 DEBUG [59e6b60f6143:41405 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-11T03:15:34,713 DEBUG [59e6b60f6143:41405 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-11T03:15:34,713 DEBUG [59e6b60f6143:41405 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-11T03:15:34,713 DEBUG [59e6b60f6143:41405 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-11T03:15:34,713 INFO [59e6b60f6143:41405 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-11T03:15:34,714 INFO [59e6b60f6143:41405 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-11T03:15:34,714 INFO [59e6b60f6143:41405 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-11T03:15:34,714 DEBUG [59e6b60f6143:41405 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-11T03:15:34,714 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,716 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 59e6b60f6143,41049,1731294933989, state=OPENING 2024-11-11T03:15:34,718 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-11-11T03:15:34,719 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,719 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,719 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,719 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:34,720 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-11T03:15:34,720 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:34,720 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:34,720 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:34,720 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=59e6b60f6143,41049,1731294933989}] 2024-11-11T03:15:34,720 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:34,874 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-11T03:15:34,876 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55363, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-11T03:15:34,881 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-11-11T03:15:34,881 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-11T03:15:34,883 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=59e6b60f6143%2C41049%2C1731294933989.meta, suffix=.meta, logDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,41049,1731294933989, archiveDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/oldWALs, maxLogs=32 2024-11-11T03:15:34,884 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 59e6b60f6143%2C41049%2C1731294933989.meta.1731294934884.meta 2024-11-11T03:15:34,892 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/WALs/59e6b60f6143,41049,1731294933989/59e6b60f6143%2C41049%2C1731294933989.meta.1731294934884.meta 2024-11-11T03:15:34,897 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38055:38055),(127.0.0.1/127.0.0.1:37869:37869),(127.0.0.1/127.0.0.1:42701:42701)] 2024-11-11T03:15:34,901 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-11-11T03:15:34,901 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-11-11T03:15:34,901 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-11-11T03:15:34,901 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-11-11T03:15:34,901 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-11-11T03:15:34,901 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:34,901 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-11-11T03:15:34,902 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-11-11T03:15:34,903 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-11T03:15:34,904 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-11T03:15:34,904 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,905 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,905 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-11T03:15:34,906 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-11T03:15:34,906 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,906 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,907 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-11T03:15:34,907 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-11T03:15:34,907 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,908 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,908 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-11T03:15:34,909 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-11T03:15:34,909 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:34,909 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-11T03:15:34,910 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-11T03:15:34,910 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740 2024-11-11T03:15:34,912 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740 2024-11-11T03:15:34,913 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-11T03:15:34,913 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-11T03:15:34,914 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-11T03:15:34,915 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-11T03:15:34,916 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67639276, jitterRate=0.007903754711151123}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-11T03:15:34,916 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-11-11T03:15:34,917 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1731294934902Writing region info on filesystem at 1731294934902Initializing all the Stores at 1731294934903 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294934903Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294934903Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294934903Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731294934903Cleaning up temporary data from old regions at 1731294934913 (+10 ms)Running coprocessor post-open hooks at 1731294934917 (+4 ms)Region opened successfully at 1731294934917 2024-11-11T03:15:34,919 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1731294934874 2024-11-11T03:15:34,922 DEBUG [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-11-11T03:15:34,922 INFO [RS_OPEN_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-11-11T03:15:34,923 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,925 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 59e6b60f6143,41049,1731294933989, state=OPEN 2024-11-11T03:15:34,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-11T03:15:34,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-11T03:15:34,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-11T03:15:34,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-11T03:15:34,927 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=59e6b60f6143,41049,1731294933989 2024-11-11T03:15:34,927 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:34,927 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:34,927 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:34,927 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-11T03:15:34,931 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-11-11T03:15:34,931 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=59e6b60f6143,41049,1731294933989 in 207 msec 2024-11-11T03:15:34,935 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-11-11T03:15:34,935 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 624 msec 2024-11-11T03:15:34,936 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-11T03:15:34,936 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-11-11T03:15:34,938 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-11T03:15:34,938 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=59e6b60f6143,41049,1731294933989, seqNum=-1] 2024-11-11T03:15:34,938 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-11T03:15:34,940 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34551, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-11T03:15:34,948 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 701 msec 2024-11-11T03:15:34,948 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1731294934948, completionTime=-1 2024-11-11T03:15:34,948 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-11-11T03:15:34,948 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] assignment.AssignmentManager(1756): Joining cluster... 2024-11-11T03:15:34,950 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] assignment.AssignmentManager(1768): Number of RegionServers=3 2024-11-11T03:15:34,950 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1731294994950 2024-11-11T03:15:34,950 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1731295054950 2024-11-11T03:15:34,950 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] assignment.AssignmentManager(1775): Joined the cluster in 2 msec 2024-11-11T03:15:34,951 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41405,1731294933908-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,951 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41405,1731294933908-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,951 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41405,1731294933908-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,951 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-59e6b60f6143:41405, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,951 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,951 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:34,953 DEBUG [master/59e6b60f6143:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-11-11T03:15:34,956 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.896sec 2024-11-11T03:15:34,956 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-11-11T03:15:34,956 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-11-11T03:15:34,956 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-11-11T03:15:34,956 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-11-11T03:15:34,956 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-11-11T03:15:34,956 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41405,1731294933908-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-11T03:15:34,956 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41405,1731294933908-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-11-11T03:15:34,958 DEBUG [master/59e6b60f6143:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-11-11T03:15:34,958 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-11-11T03:15:34,959 INFO [master/59e6b60f6143:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=59e6b60f6143,41405,1731294933908-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-11T03:15:35,039 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@61e12713, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-11T03:15:35,039 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 59e6b60f6143,41405,-1 for getting cluster id 2024-11-11T03:15:35,039 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-11-11T03:15:35,040 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'b16762e3-d680-4989-9837-cbc94cfa07e2' 2024-11-11T03:15:35,041 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-11-11T03:15:35,041 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "b16762e3-d680-4989-9837-cbc94cfa07e2" 2024-11-11T03:15:35,041 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@76e7acfe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-11T03:15:35,041 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [59e6b60f6143,41405,-1] 2024-11-11T03:15:35,041 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-11-11T03:15:35,042 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:35,043 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51514, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-11-11T03:15:35,044 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@cb57756, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-11T03:15:35,045 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-11T03:15:35,046 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=59e6b60f6143,41049,1731294933989, seqNum=-1] 2024-11-11T03:15:35,046 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-11T03:15:35,048 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34032, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-11T03:15:35,050 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=59e6b60f6143,41405,1731294933908 2024-11-11T03:15:35,051 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-11-11T03:15:35,052 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.AsyncConnectionImpl(321): The fetched master address is 59e6b60f6143,41405,1731294933908 2024-11-11T03:15:35,052 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@2a76b4f 2024-11-11T03:15:35,053 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-11-11T03:15:35,054 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51524, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-11-11T03:15:35,055 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-11T03:15:35,057 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-11-11T03:15:35,059 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-11-11T03:15:35,059 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:35,059 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-11-11T03:15:35,060 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-11T03:15:35,061 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-11T03:15:35,071 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741837_1013 (size=392) 2024-11-11T03:15:35,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741837_1013 (size=392) 2024-11-11T03:15:35,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741837_1013 (size=392) 2024-11-11T03:15:35,075 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 12f088d4ef10d1fe3520b9d0d0fc8500, NAME => 'TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605 2024-11-11T03:15:35,085 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741838_1014 (size=51) 2024-11-11T03:15:35,086 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741838_1014 (size=51) 2024-11-11T03:15:35,086 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741838_1014 (size=51) 2024-11-11T03:15:35,087 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:35,087 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 12f088d4ef10d1fe3520b9d0d0fc8500, disabling compactions & flushes 2024-11-11T03:15:35,087 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,087 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,088 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. after waiting 0 ms 2024-11-11T03:15:35,088 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,088 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,088 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 12f088d4ef10d1fe3520b9d0d0fc8500: Waiting for close lock at 1731294935087Disabling compacts and flushes for region at 1731294935087Disabling writes for close at 1731294935088 (+1 ms)Writing region close event to WAL at 1731294935088Closed at 1731294935088 2024-11-11T03:15:35,090 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-11-11T03:15:35,090 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1731294935090"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1731294935090"}]},"ts":"1731294935090"} 2024-11-11T03:15:35,093 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-11-11T03:15:35,095 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-11-11T03:15:35,095 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731294935095"}]},"ts":"1731294935095"} 2024-11-11T03:15:35,098 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-11-11T03:15:35,098 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {59e6b60f6143=0} racks are {/default-rack=0} 2024-11-11T03:15:35,099 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-11T03:15:35,099 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-11T03:15:35,099 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-11T03:15:35,099 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-11T03:15:35,099 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-11T03:15:35,099 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-11T03:15:35,099 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-11T03:15:35,099 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-11T03:15:35,099 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-11T03:15:35,099 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-11T03:15:35,100 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=12f088d4ef10d1fe3520b9d0d0fc8500, ASSIGN}] 2024-11-11T03:15:35,102 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=12f088d4ef10d1fe3520b9d0d0fc8500, ASSIGN 2024-11-11T03:15:35,103 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=12f088d4ef10d1fe3520b9d0d0fc8500, ASSIGN; state=OFFLINE, location=59e6b60f6143,41097,1731294933960; forceNewPlan=false, retain=false 2024-11-11T03:15:35,168 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-11T03:15:35,254 INFO [59e6b60f6143:41405 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-11-11T03:15:35,254 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=12f088d4ef10d1fe3520b9d0d0fc8500, regionState=OPENING, regionLocation=59e6b60f6143,41097,1731294933960 2024-11-11T03:15:35,258 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=12f088d4ef10d1fe3520b9d0d0fc8500, ASSIGN because future has completed 2024-11-11T03:15:35,259 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 12f088d4ef10d1fe3520b9d0d0fc8500, server=59e6b60f6143,41097,1731294933960}] 2024-11-11T03:15:35,379 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-11T03:15:35,413 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-11T03:15:35,415 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52417, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-11T03:15:35,420 INFO [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,421 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 12f088d4ef10d1fe3520b9d0d0fc8500, NAME => 'TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500.', STARTKEY => '', ENDKEY => ''} 2024-11-11T03:15:35,421 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,421 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-11T03:15:35,422 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,422 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,424 INFO [StoreOpener-12f088d4ef10d1fe3520b9d0d0fc8500-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,425 INFO [StoreOpener-12f088d4ef10d1fe3520b9d0d0fc8500-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 12f088d4ef10d1fe3520b9d0d0fc8500 columnFamilyName cf 2024-11-11T03:15:35,425 DEBUG [StoreOpener-12f088d4ef10d1fe3520b9d0d0fc8500-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-11T03:15:35,426 INFO [StoreOpener-12f088d4ef10d1fe3520b9d0d0fc8500-1 {}] regionserver.HStore(327): Store=12f088d4ef10d1fe3520b9d0d0fc8500/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-11T03:15:35,426 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,427 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,428 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,428 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,428 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,430 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,433 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-11T03:15:35,434 INFO [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 12f088d4ef10d1fe3520b9d0d0fc8500; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68477944, jitterRate=0.02040088176727295}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-11-11T03:15:35,434 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:35,435 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 12f088d4ef10d1fe3520b9d0d0fc8500: Running coprocessor pre-open hook at 1731294935422Writing region info on filesystem at 1731294935422Initializing all the Stores at 1731294935423 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731294935423Cleaning up temporary data from old regions at 1731294935428 (+5 ms)Running coprocessor post-open hooks at 1731294935434 (+6 ms)Region opened successfully at 1731294935435 (+1 ms) 2024-11-11T03:15:35,437 INFO [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500., pid=6, masterSystemTime=1731294935413 2024-11-11T03:15:35,441 DEBUG [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,441 INFO [RS_OPEN_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,442 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=12f088d4ef10d1fe3520b9d0d0fc8500, regionState=OPEN, openSeqNum=2, regionLocation=59e6b60f6143,41097,1731294933960 2024-11-11T03:15:35,445 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 12f088d4ef10d1fe3520b9d0d0fc8500, server=59e6b60f6143,41097,1731294933960 because future has completed 2024-11-11T03:15:35,451 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-11-11T03:15:35,451 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 12f088d4ef10d1fe3520b9d0d0fc8500, server=59e6b60f6143,41097,1731294933960 in 188 msec 2024-11-11T03:15:35,456 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-11-11T03:15:35,456 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=12f088d4ef10d1fe3520b9d0d0fc8500, ASSIGN in 352 msec 2024-11-11T03:15:35,457 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-11-11T03:15:35,458 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731294935457"}]},"ts":"1731294935457"} 2024-11-11T03:15:35,461 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-11-11T03:15:35,462 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-11-11T03:15:35,465 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 407 msec 2024-11-11T03:15:35,583 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-11-11T03:15:35,589 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-11T03:15:35,632 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-11T03:15:35,633 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-11T03:15:35,688 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-11T03:15:35,688 INFO [RPCClient-NioEventLoopGroup-6-8 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-11-11T03:15:35,688 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-11-11T03:15:35,689 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-11T03:15:35,692 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-11-11T03:15:35,692 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-11T03:15:35,692 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-11-11T03:15:35,695 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500., hostname=59e6b60f6143,41097,1731294933960, seqNum=2] 2024-11-11T03:15:35,695 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-11T03:15:35,697 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43704, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-11T03:15:35,700 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-11-11T03:15:35,702 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-11-11T03:15:35,703 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-11T03:15:35,704 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-11-11T03:15:35,705 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-11-11T03:15:35,705 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-11-11T03:15:35,808 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-11T03:15:35,860 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=41097 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-11-11T03:15:35,860 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,860 INFO [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 12f088d4ef10d1fe3520b9d0d0fc8500 1/1 column families, dataSize=32 B heapSize=360 B 2024-11-11T03:15:35,880 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500/.tmp/cf/45785660a6914004b525435428aadbf6 is 36, key is row/cf:cq/1731294935698/Put/seqid=0 2024-11-11T03:15:35,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741839_1015 (size=4787) 2024-11-11T03:15:35,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741839_1015 (size=4787) 2024-11-11T03:15:35,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741839_1015 (size=4787) 2024-11-11T03:15:35,890 INFO [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500/.tmp/cf/45785660a6914004b525435428aadbf6 2024-11-11T03:15:35,900 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500/.tmp/cf/45785660a6914004b525435428aadbf6 as hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500/cf/45785660a6914004b525435428aadbf6 2024-11-11T03:15:35,907 INFO [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500/cf/45785660a6914004b525435428aadbf6, entries=1, sequenceid=5, filesize=4.7 K 2024-11-11T03:15:35,909 INFO [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 12f088d4ef10d1fe3520b9d0d0fc8500 in 49ms, sequenceid=5, compaction requested=false 2024-11-11T03:15:35,909 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 12f088d4ef10d1fe3520b9d0d0fc8500: 2024-11-11T03:15:35,909 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:35,910 DEBUG [RS_FLUSH_OPERATIONS-regionserver/59e6b60f6143:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-11-11T03:15:35,910 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-11-11T03:15:35,916 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-11-11T03:15:35,917 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 207 msec 2024-11-11T03:15:35,920 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 217 msec 2024-11-11T03:15:36,019 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41405 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-11T03:15:36,019 INFO [RPCClient-NioEventLoopGroup-6-8 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-11-11T03:15:36,023 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-11-11T03:15:36,023 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-11T03:15:36,023 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:36,023 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,024 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,024 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-11-11T03:15:36,024 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-11-11T03:15:36,024 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1012825180, stopped=false 2024-11-11T03:15:36,024 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=59e6b60f6143,41405,1731294933908 2024-11-11T03:15:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-11T03:15:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:36,026 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-11T03:15:36,026 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-11T03:15:36,027 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:36,027 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,027 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:36,027 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:36,027 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '59e6b60f6143,41097,1731294933960' ***** 2024-11-11T03:15:36,027 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:36,027 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-11T03:15:36,027 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '59e6b60f6143,41049,1731294933989' ***** 2024-11-11T03:15:36,028 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-11T03:15:36,028 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-11T03:15:36,028 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-11T03:15:36,028 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '59e6b60f6143,35667,1731294934017' ***** 2024-11-11T03:15:36,028 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-11T03:15:36,028 INFO [RS:0;59e6b60f6143:41097 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-11T03:15:36,028 INFO [RS:0;59e6b60f6143:41097 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-11T03:15:36,028 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-11T03:15:36,028 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(3091): Received CLOSE for 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:36,028 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-11T03:15:36,028 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(959): stopping server 59e6b60f6143,41049,1731294933989 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-11T03:15:36,029 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;59e6b60f6143:41049. 2024-11-11T03:15:36,029 INFO [RS:2;59e6b60f6143:35667 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-11T03:15:36,029 DEBUG [RS:1;59e6b60f6143:41049 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:36,029 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-11T03:15:36,029 INFO [RS:2;59e6b60f6143:35667 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-11T03:15:36,029 DEBUG [RS:1;59e6b60f6143:41049 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,029 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(959): stopping server 59e6b60f6143,35667,1731294934017 2024-11-11T03:15:36,029 INFO [RS:2;59e6b60f6143:35667 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-11T03:15:36,029 INFO [RS:2;59e6b60f6143:35667 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;59e6b60f6143:35667. 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-11T03:15:36,029 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-11-11T03:15:36,029 DEBUG [RS:2;59e6b60f6143:35667 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:36,030 DEBUG [RS:2;59e6b60f6143:35667 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,030 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(959): stopping server 59e6b60f6143,41097,1731294933960 2024-11-11T03:15:36,030 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(976): stopping server 59e6b60f6143,35667,1731294934017; all regions closed. 2024-11-11T03:15:36,030 INFO [RS:0;59e6b60f6143:41097 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-11T03:15:36,030 INFO [RS:0;59e6b60f6143:41097 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;59e6b60f6143:41097. 2024-11-11T03:15:36,030 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 12f088d4ef10d1fe3520b9d0d0fc8500, disabling compactions & flushes 2024-11-11T03:15:36,030 INFO [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:36,030 DEBUG [RS:0;59e6b60f6143:41097 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-11T03:15:36,030 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:36,030 DEBUG [RS:0;59e6b60f6143:41097 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,030 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. after waiting 0 ms 2024-11-11T03:15:36,030 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:36,030 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-11T03:15:36,030 DEBUG [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1325): Online Regions={12f088d4ef10d1fe3520b9d0d0fc8500=TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500.} 2024-11-11T03:15:36,030 DEBUG [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1351): Waiting on 12f088d4ef10d1fe3520b9d0d0fc8500 2024-11-11T03:15:36,032 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,032 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-11T03:15:36,032 DEBUG [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-11-11T03:15:36,032 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,032 DEBUG [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-11-11T03:15:36,032 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,032 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-11T03:15:36,032 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,032 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-11T03:15:36,032 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-11T03:15:36,032 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,032 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-11T03:15:36,032 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-11T03:15:36,032 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-11-11T03:15:36,035 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741835_1011 (size=93) 2024-11-11T03:15:36,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741835_1011 (size=93) 2024-11-11T03:15:36,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741835_1011 (size=93) 2024-11-11T03:15:36,038 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/default/TestHBaseWalOnEC/12f088d4ef10d1fe3520b9d0d0fc8500/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-11-11T03:15:36,039 INFO [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:36,039 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 12f088d4ef10d1fe3520b9d0d0fc8500: Waiting for close lock at 1731294936030Running coprocessor pre-close hooks at 1731294936030Disabling compacts and flushes for region at 1731294936030Disabling writes for close at 1731294936030Writing region close event to WAL at 1731294936032 (+2 ms)Running coprocessor post-close hooks at 1731294936039 (+7 ms)Closed at 1731294936039 2024-11-11T03:15:36,040 DEBUG [RS_CLOSE_REGION-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500. 2024-11-11T03:15:36,040 DEBUG [RS:2;59e6b60f6143:35667 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/oldWALs 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 59e6b60f6143%2C35667%2C1731294934017:(num 1731294934537) 2024-11-11T03:15:36,040 DEBUG [RS:2;59e6b60f6143:35667 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] hbase.ChoreService(370): Chore service for: regionserver/59e6b60f6143:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-11T03:15:36,040 INFO [regionserver/59e6b60f6143:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-11T03:15:36,040 INFO [RS:2;59e6b60f6143:35667 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:35667 2024-11-11T03:15:36,042 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/59e6b60f6143,35667,1731294934017 2024-11-11T03:15:36,042 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-11T03:15:36,042 INFO [RS:2;59e6b60f6143:35667 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-11T03:15:36,044 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [59e6b60f6143,35667,1731294934017] 2024-11-11T03:15:36,046 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/59e6b60f6143,35667,1731294934017 already deleted, retry=false 2024-11-11T03:15:36,046 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 59e6b60f6143,35667,1731294934017 expired; onlineServers=2 2024-11-11T03:15:36,054 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/info/d6e34ed484684b8d814a9cbe40a5d2e3 is 153, key is TestHBaseWalOnEC,,1731294935055.12f088d4ef10d1fe3520b9d0d0fc8500./info:regioninfo/1731294935441/Put/seqid=0 2024-11-11T03:15:36,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741840_1016 (size=6637) 2024-11-11T03:15:36,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741840_1016 (size=6637) 2024-11-11T03:15:36,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741840_1016 (size=6637) 2024-11-11T03:15:36,063 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/info/d6e34ed484684b8d814a9cbe40a5d2e3 2024-11-11T03:15:36,087 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/ns/25910166f37842ce8c1c2ec4ede202ef is 43, key is default/ns:d/1731294934941/Put/seqid=0 2024-11-11T03:15:36,093 INFO [regionserver/59e6b60f6143:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:36,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741841_1017 (size=5153) 2024-11-11T03:15:36,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741841_1017 (size=5153) 2024-11-11T03:15:36,095 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741841_1017 (size=5153) 2024-11-11T03:15:36,095 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/ns/25910166f37842ce8c1c2ec4ede202ef 2024-11-11T03:15:36,098 INFO [regionserver/59e6b60f6143:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:36,098 INFO [regionserver/59e6b60f6143:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:36,121 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/table/a4145793a97b4bd883ff21689f65692a is 52, key is TestHBaseWalOnEC/table:state/1731294935457/Put/seqid=0 2024-11-11T03:15:36,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741842_1018 (size=5249) 2024-11-11T03:15:36,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741842_1018 (size=5249) 2024-11-11T03:15:36,129 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741842_1018 (size=5249) 2024-11-11T03:15:36,129 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/table/a4145793a97b4bd883ff21689f65692a 2024-11-11T03:15:36,137 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/info/d6e34ed484684b8d814a9cbe40a5d2e3 as hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/info/d6e34ed484684b8d814a9cbe40a5d2e3 2024-11-11T03:15:36,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:36,144 INFO [RS:2;59e6b60f6143:35667 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-11T03:15:36,144 INFO [RS:2;59e6b60f6143:35667 {}] regionserver.HRegionServer(1031): Exiting; stopping=59e6b60f6143,35667,1731294934017; zookeeper connection closed. 2024-11-11T03:15:36,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35667-0x10105061e0f0003, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:36,145 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@409b07dc {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@409b07dc 2024-11-11T03:15:36,146 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/info/d6e34ed484684b8d814a9cbe40a5d2e3, entries=10, sequenceid=11, filesize=6.5 K 2024-11-11T03:15:36,148 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/ns/25910166f37842ce8c1c2ec4ede202ef as hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/ns/25910166f37842ce8c1c2ec4ede202ef 2024-11-11T03:15:36,155 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/ns/25910166f37842ce8c1c2ec4ede202ef, entries=2, sequenceid=11, filesize=5.0 K 2024-11-11T03:15:36,156 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/.tmp/table/a4145793a97b4bd883ff21689f65692a as hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/table/a4145793a97b4bd883ff21689f65692a 2024-11-11T03:15:36,163 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/table/a4145793a97b4bd883ff21689f65692a, entries=2, sequenceid=11, filesize=5.1 K 2024-11-11T03:15:36,165 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 132ms, sequenceid=11, compaction requested=false 2024-11-11T03:15:36,170 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-11-11T03:15:36,171 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-11T03:15:36,171 INFO [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-11T03:15:36,171 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731294936032Running coprocessor pre-close hooks at 1731294936032Disabling compacts and flushes for region at 1731294936032Disabling writes for close at 1731294936032Obtaining lock to block concurrent updates at 1731294936032Preparing flush snapshotting stores in 1588230740 at 1731294936032Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1731294936033 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1731294936034 (+1 ms)Flushing 1588230740/info: creating writer at 1731294936034Flushing 1588230740/info: appending metadata at 1731294936054 (+20 ms)Flushing 1588230740/info: closing flushed file at 1731294936054Flushing 1588230740/ns: creating writer at 1731294936070 (+16 ms)Flushing 1588230740/ns: appending metadata at 1731294936087 (+17 ms)Flushing 1588230740/ns: closing flushed file at 1731294936087Flushing 1588230740/table: creating writer at 1731294936103 (+16 ms)Flushing 1588230740/table: appending metadata at 1731294936120 (+17 ms)Flushing 1588230740/table: closing flushed file at 1731294936120Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1010947b: reopening flushed file at 1731294936136 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4be46117: reopening flushed file at 1731294936147 (+11 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@30a5811f: reopening flushed file at 1731294936155 (+8 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 132ms, sequenceid=11, compaction requested=false at 1731294936165 (+10 ms)Writing region close event to WAL at 1731294936166 (+1 ms)Running coprocessor post-close hooks at 1731294936171 (+5 ms)Closed at 1731294936171 2024-11-11T03:15:36,171 DEBUG [RS_CLOSE_META-regionserver/59e6b60f6143:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-11-11T03:15:36,231 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(976): stopping server 59e6b60f6143,41097,1731294933960; all regions closed. 2024-11-11T03:15:36,231 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,231 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,232 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,232 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,232 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,232 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(976): stopping server 59e6b60f6143,41049,1731294933989; all regions closed. 2024-11-11T03:15:36,233 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,233 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,233 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,233 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,233 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,235 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741834_1010 (size=1298) 2024-11-11T03:15:36,235 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741834_1010 (size=1298) 2024-11-11T03:15:36,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741836_1012 (size=2751) 2024-11-11T03:15:36,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741836_1012 (size=2751) 2024-11-11T03:15:36,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741836_1012 (size=2751) 2024-11-11T03:15:36,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741834_1010 (size=1298) 2024-11-11T03:15:36,241 DEBUG [RS:0;59e6b60f6143:41097 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/oldWALs 2024-11-11T03:15:36,241 INFO [RS:0;59e6b60f6143:41097 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 59e6b60f6143%2C41097%2C1731294933960:(num 1731294934536) 2024-11-11T03:15:36,241 DEBUG [RS:0;59e6b60f6143:41097 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,241 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:36,241 INFO [RS:0;59e6b60f6143:41097 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-11T03:15:36,242 DEBUG [RS:1;59e6b60f6143:41049 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/oldWALs 2024-11-11T03:15:36,242 INFO [RS:1;59e6b60f6143:41049 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 59e6b60f6143%2C41049%2C1731294933989.meta:.meta(num 1731294934884) 2024-11-11T03:15:36,242 INFO [RS:0;59e6b60f6143:41097 {}] hbase.ChoreService(370): Chore service for: regionserver/59e6b60f6143:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-11T03:15:36,242 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,242 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-11T03:15:36,242 INFO [regionserver/59e6b60f6143:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-11T03:15:36,242 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-11T03:15:36,242 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-11T03:15:36,242 INFO [RS:0;59e6b60f6143:41097 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-11T03:15:36,243 INFO [RS:0;59e6b60f6143:41097 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41097 2024-11-11T03:15:36,243 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,244 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,244 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,244 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,245 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/59e6b60f6143,41097,1731294933960 2024-11-11T03:15:36,245 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-11T03:15:36,245 INFO [RS:0;59e6b60f6143:41097 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-11T03:15:36,246 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741833_1009 (size=93) 2024-11-11T03:15:36,247 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [59e6b60f6143,41097,1731294933960] 2024-11-11T03:15:36,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741833_1009 (size=93) 2024-11-11T03:15:36,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741833_1009 (size=93) 2024-11-11T03:15:36,248 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/59e6b60f6143,41097,1731294933960 already deleted, retry=false 2024-11-11T03:15:36,249 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 59e6b60f6143,41097,1731294933960 expired; onlineServers=1 2024-11-11T03:15:36,251 DEBUG [RS:1;59e6b60f6143:41049 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/oldWALs 2024-11-11T03:15:36,251 INFO [RS:1;59e6b60f6143:41049 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 59e6b60f6143%2C41049%2C1731294933989:(num 1731294934532) 2024-11-11T03:15:36,251 DEBUG [RS:1;59e6b60f6143:41049 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-11T03:15:36,251 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.LeaseManager(133): Closed leases 2024-11-11T03:15:36,251 INFO [RS:1;59e6b60f6143:41049 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-11T03:15:36,251 INFO [RS:1;59e6b60f6143:41049 {}] hbase.ChoreService(370): Chore service for: regionserver/59e6b60f6143:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-11T03:15:36,251 INFO [RS:1;59e6b60f6143:41049 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-11T03:15:36,251 INFO [regionserver/59e6b60f6143:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-11T03:15:36,251 INFO [RS:1;59e6b60f6143:41049 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41049 2024-11-11T03:15:36,254 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/59e6b60f6143,41049,1731294933989 2024-11-11T03:15:36,254 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-11T03:15:36,254 INFO [RS:1;59e6b60f6143:41049 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-11T03:15:36,256 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [59e6b60f6143,41049,1731294933989] 2024-11-11T03:15:36,258 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/59e6b60f6143,41049,1731294933989 already deleted, retry=false 2024-11-11T03:15:36,258 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 59e6b60f6143,41049,1731294933989 expired; onlineServers=0 2024-11-11T03:15:36,258 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '59e6b60f6143,41405,1731294933908' ***** 2024-11-11T03:15:36,258 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-11-11T03:15:36,258 INFO [M:0;59e6b60f6143:41405 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-11T03:15:36,258 INFO [M:0;59e6b60f6143:41405 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-11T03:15:36,258 DEBUG [M:0;59e6b60f6143:41405 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-11-11T03:15:36,258 DEBUG [M:0;59e6b60f6143:41405 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-11-11T03:15:36,258 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-11-11T03:15:36,258 DEBUG [master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.small.0-1731294934256 {}] cleaner.HFileCleaner(306): Exit Thread[master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.small.0-1731294934256,5,FailOnTimeoutGroup] 2024-11-11T03:15:36,258 INFO [M:0;59e6b60f6143:41405 {}] hbase.ChoreService(370): Chore service for: master/59e6b60f6143:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-11-11T03:15:36,258 DEBUG [master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.large.0-1731294934252 {}] cleaner.HFileCleaner(306): Exit Thread[master/59e6b60f6143:0:becomeActiveMaster-HFileCleaner.large.0-1731294934252,5,FailOnTimeoutGroup] 2024-11-11T03:15:36,259 INFO [M:0;59e6b60f6143:41405 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-11T03:15:36,259 DEBUG [M:0;59e6b60f6143:41405 {}] master.HMaster(1795): Stopping service threads 2024-11-11T03:15:36,259 INFO [M:0;59e6b60f6143:41405 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-11-11T03:15:36,259 INFO [M:0;59e6b60f6143:41405 {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-11T03:15:36,259 INFO [M:0;59e6b60f6143:41405 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-11-11T03:15:36,259 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-11-11T03:15:36,260 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-11-11T03:15:36,260 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-11T03:15:36,260 DEBUG [M:0;59e6b60f6143:41405 {}] zookeeper.ZKUtil(347): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-11-11T03:15:36,260 WARN [M:0;59e6b60f6143:41405 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-11-11T03:15:36,261 INFO [M:0;59e6b60f6143:41405 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/.lastflushedseqids 2024-11-11T03:15:36,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741843_1019 (size=127) 2024-11-11T03:15:36,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741843_1019 (size=127) 2024-11-11T03:15:36,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741843_1019 (size=127) 2024-11-11T03:15:36,271 INFO [M:0;59e6b60f6143:41405 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-11-11T03:15:36,272 INFO [M:0;59e6b60f6143:41405 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-11-11T03:15:36,272 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-11T03:15:36,272 INFO [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:36,272 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:36,272 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-11T03:15:36,272 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:36,272 INFO [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.82 KB heapSize=34.11 KB 2024-11-11T03:15:36,300 DEBUG [M:0;59e6b60f6143:41405 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e4c2bc5f6e05481aa88225307a0d7277 is 82, key is hbase:meta,,1/info:regioninfo/1731294934923/Put/seqid=0 2024-11-11T03:15:36,309 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741844_1020 (size=5672) 2024-11-11T03:15:36,309 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741844_1020 (size=5672) 2024-11-11T03:15:36,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741844_1020 (size=5672) 2024-11-11T03:15:36,310 INFO [M:0;59e6b60f6143:41405 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e4c2bc5f6e05481aa88225307a0d7277 2024-11-11T03:15:36,344 DEBUG [M:0;59e6b60f6143:41405 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/aec15c1a7c4949bcbb3fa7b8c6b10645 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1731294935464/Put/seqid=0 2024-11-11T03:15:36,348 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:36,348 INFO [RS:0;59e6b60f6143:41097 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-11T03:15:36,348 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41097-0x10105061e0f0001, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:36,348 INFO [RS:0;59e6b60f6143:41097 {}] regionserver.HRegionServer(1031): Exiting; stopping=59e6b60f6143,41097,1731294933960; zookeeper connection closed. 2024-11-11T03:15:36,348 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@84f6f4e {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@84f6f4e 2024-11-11T03:15:36,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741845_1021 (size=6438) 2024-11-11T03:15:36,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741845_1021 (size=6438) 2024-11-11T03:15:36,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741845_1021 (size=6438) 2024-11-11T03:15:36,355 INFO [M:0;59e6b60f6143:41405 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/aec15c1a7c4949bcbb3fa7b8c6b10645 2024-11-11T03:15:36,356 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:36,356 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10105061e0f0002, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:36,356 INFO [RS:1;59e6b60f6143:41049 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-11T03:15:36,356 INFO [RS:1;59e6b60f6143:41049 {}] regionserver.HRegionServer(1031): Exiting; stopping=59e6b60f6143,41049,1731294933989; zookeeper connection closed. 2024-11-11T03:15:36,357 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@6bf7ba16 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@6bf7ba16 2024-11-11T03:15:36,357 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-11-11T03:15:36,384 DEBUG [M:0;59e6b60f6143:41405 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/46a329a470b24edab057b1a106a5d1fa is 69, key is 59e6b60f6143,35667,1731294934017/rs:state/1731294934364/Put/seqid=0 2024-11-11T03:15:36,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741846_1022 (size=5294) 2024-11-11T03:15:36,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741846_1022 (size=5294) 2024-11-11T03:15:36,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741846_1022 (size=5294) 2024-11-11T03:15:36,393 INFO [M:0;59e6b60f6143:41405 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/46a329a470b24edab057b1a106a5d1fa 2024-11-11T03:15:36,400 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e4c2bc5f6e05481aa88225307a0d7277 as hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/e4c2bc5f6e05481aa88225307a0d7277 2024-11-11T03:15:36,407 INFO [M:0;59e6b60f6143:41405 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/e4c2bc5f6e05481aa88225307a0d7277, entries=8, sequenceid=72, filesize=5.5 K 2024-11-11T03:15:36,409 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/aec15c1a7c4949bcbb3fa7b8c6b10645 as hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/aec15c1a7c4949bcbb3fa7b8c6b10645 2024-11-11T03:15:36,416 INFO [M:0;59e6b60f6143:41405 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/aec15c1a7c4949bcbb3fa7b8c6b10645, entries=8, sequenceid=72, filesize=6.3 K 2024-11-11T03:15:36,417 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/46a329a470b24edab057b1a106a5d1fa as hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/46a329a470b24edab057b1a106a5d1fa 2024-11-11T03:15:36,424 INFO [M:0;59e6b60f6143:41405 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38147/user/jenkins/test-data/ffe3919c-2258-5ed9-b44e-4b9002179605/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/46a329a470b24edab057b1a106a5d1fa, entries=3, sequenceid=72, filesize=5.2 K 2024-11-11T03:15:36,426 INFO [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.82 KB/27462, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 154ms, sequenceid=72, compaction requested=false 2024-11-11T03:15:36,428 INFO [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-11T03:15:36,428 DEBUG [M:0;59e6b60f6143:41405 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731294936272Disabling compacts and flushes for region at 1731294936272Disabling writes for close at 1731294936272Obtaining lock to block concurrent updates at 1731294936272Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1731294936272Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27462, getHeapSize=34864, getOffHeapSize=0, getCellsCount=85 at 1731294936273 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1731294936274 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1731294936274Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1731294936299 (+25 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1731294936299Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1731294936318 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1731294936344 (+26 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1731294936344Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1731294936362 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1731294936383 (+21 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1731294936383Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@717b03f7: reopening flushed file at 1731294936399 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@60df730f: reopening flushed file at 1731294936407 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@f432a1c: reopening flushed file at 1731294936416 (+9 ms)Finished flush of dataSize ~26.82 KB/27462, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 154ms, sequenceid=72, compaction requested=false at 1731294936426 (+10 ms)Writing region close event to WAL at 1731294936427 (+1 ms)Closed at 1731294936427 2024-11-11T03:15:36,428 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,428 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,428 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,429 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,429 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-11T03:15:36,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41783 is added to blk_1073741830_1006 (size=32665) 2024-11-11T03:15:36,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44627 is added to blk_1073741830_1006 (size=32665) 2024-11-11T03:15:36,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41935 is added to blk_1073741830_1006 (size=32665) 2024-11-11T03:15:36,833 INFO [M:0;59e6b60f6143:41405 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-11-11T03:15:36,833 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-11T03:15:36,834 INFO [M:0;59e6b60f6143:41405 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41405 2024-11-11T03:15:36,834 INFO [M:0;59e6b60f6143:41405 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-11T03:15:36,937 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:36,937 INFO [M:0;59e6b60f6143:41405 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-11T03:15:36,937 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41405-0x10105061e0f0000, quorum=127.0.0.1:60206, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-11T03:15:36,940 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@25c162fe{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:36,940 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1da1746e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-11T03:15:36,940 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-11T03:15:36,940 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@739551bc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-11T03:15:36,940 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7adc0795{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir/,STOPPED} 2024-11-11T03:15:36,942 WARN [BP-1808564834-172.17.0.2-1731294932919 heartbeating to localhost/127.0.0.1:38147 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-11T03:15:36,942 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-11T03:15:36,942 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-11T03:15:36,942 WARN [BP-1808564834-172.17.0.2-1731294932919 heartbeating to localhost/127.0.0.1:38147 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1808564834-172.17.0.2-1731294932919 (Datanode Uuid b1a66a44-013b-4763-aedf-4dd4bc268fa6) service to localhost/127.0.0.1:38147 2024-11-11T03:15:36,943 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data5/current/BP-1808564834-172.17.0.2-1731294932919 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:36,943 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data6/current/BP-1808564834-172.17.0.2-1731294932919 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:36,943 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-11T03:15:36,945 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@609afea4{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:36,946 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@527d4f04{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-11T03:15:36,946 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-11T03:15:36,946 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6f8ad177{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-11T03:15:36,946 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@69cb0b1f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir/,STOPPED} 2024-11-11T03:15:36,947 WARN [BP-1808564834-172.17.0.2-1731294932919 heartbeating to localhost/127.0.0.1:38147 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-11T03:15:36,947 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-11T03:15:36,947 WARN [BP-1808564834-172.17.0.2-1731294932919 heartbeating to localhost/127.0.0.1:38147 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1808564834-172.17.0.2-1731294932919 (Datanode Uuid d7bfa18f-a190-478f-a4dc-3c02c56ba7f4) service to localhost/127.0.0.1:38147 2024-11-11T03:15:36,947 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-11T03:15:36,948 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data4/current/BP-1808564834-172.17.0.2-1731294932919 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:36,948 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data3/current/BP-1808564834-172.17.0.2-1731294932919 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:36,948 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-11T03:15:36,950 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@14abb266{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-11T03:15:36,951 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@348ccaab{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-11T03:15:36,951 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-11T03:15:36,951 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7c2dd4e6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-11T03:15:36,951 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5a6744cf{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir/,STOPPED} 2024-11-11T03:15:36,952 WARN [BP-1808564834-172.17.0.2-1731294932919 heartbeating to localhost/127.0.0.1:38147 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-11T03:15:36,952 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-11T03:15:36,952 WARN [BP-1808564834-172.17.0.2-1731294932919 heartbeating to localhost/127.0.0.1:38147 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1808564834-172.17.0.2-1731294932919 (Datanode Uuid e2f14b05-80b8-45e2-8c6a-1a5d8d53a212) service to localhost/127.0.0.1:38147 2024-11-11T03:15:36,952 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-11T03:15:36,953 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data1/current/BP-1808564834-172.17.0.2-1731294932919 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:36,953 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/cluster_58e5e490-189b-3c7a-6176-c011752e27f4/data/data2/current/BP-1808564834-172.17.0.2-1731294932919 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-11T03:15:36,953 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-11T03:15:36,959 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3b25f894{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-11T03:15:36,960 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@53797dc3{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-11T03:15:36,960 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-11T03:15:36,960 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@715f09c8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-11T03:15:36,960 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@49b2b984{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/13d66f5c-36ed-f9c6-39fc-d67fbb00b484/hadoop.log.dir/,STOPPED} 2024-11-11T03:15:36,971 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-11-11T03:15:36,995 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-11-11T03:15:37,003 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=151 (was 89) - Thread LEAK? -, OpenFileDescriptor=521 (was 441) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=60 (was 47) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=7842 (was 8029)