2024-11-24 19:45:10,433 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-11-24 19:45:10,451 main DEBUG Took 0.015143 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-11-24 19:45:10,451 main DEBUG PluginManager 'Core' found 129 plugins 2024-11-24 19:45:10,452 main DEBUG PluginManager 'Level' found 0 plugins 2024-11-24 19:45:10,453 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-11-24 19:45:10,455 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,464 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-11-24 19:45:10,483 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,485 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,485 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,486 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,486 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,487 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,488 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,488 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,489 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,489 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,490 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,490 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,491 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,491 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,491 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,492 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,492 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,492 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,493 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,493 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,494 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,494 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,495 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,495 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-24 19:45:10,495 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,496 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-11-24 19:45:10,497 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-24 19:45:10,498 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-11-24 19:45:10,501 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-11-24 19:45:10,501 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-11-24 19:45:10,503 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-11-24 19:45:10,504 main DEBUG PluginManager 'Converter' found 47 plugins 2024-11-24 19:45:10,514 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-11-24 19:45:10,517 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-11-24 19:45:10,519 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-11-24 19:45:10,519 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-11-24 19:45:10,519 main DEBUG createAppenders(={Console}) 2024-11-24 19:45:10,520 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-11-24 19:45:10,520 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-11-24 19:45:10,521 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-11-24 19:45:10,521 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-11-24 19:45:10,521 main DEBUG OutputStream closed 2024-11-24 19:45:10,522 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-11-24 19:45:10,522 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-11-24 19:45:10,522 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-11-24 19:45:10,620 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-11-24 19:45:10,624 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-11-24 19:45:10,627 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-11-24 19:45:10,629 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-11-24 19:45:10,630 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-11-24 19:45:10,630 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-11-24 19:45:10,630 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-11-24 19:45:10,631 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-11-24 19:45:10,631 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-11-24 19:45:10,632 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-11-24 19:45:10,632 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-11-24 19:45:10,633 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-11-24 19:45:10,633 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-11-24 19:45:10,634 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-11-24 19:45:10,634 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-11-24 19:45:10,635 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-11-24 19:45:10,635 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-11-24 19:45:10,636 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-11-24 19:45:10,639 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-11-24 19:45:10,640 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-11-24 19:45:10,641 main DEBUG Shutdown hook enabled. Registering a new one. 2024-11-24 19:45:10,642 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-11-24T19:45:10,663 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-11-24 19:45:10,667 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-11-24 19:45:10,668 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-11-24T19:45:10,938 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6 2024-11-24T19:45:10,970 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620, deleteOnExit=true 2024-11-24T19:45:10,972 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/test.cache.data in system properties and HBase conf 2024-11-24T19:45:10,973 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.tmp.dir in system properties and HBase conf 2024-11-24T19:45:10,974 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir in system properties and HBase conf 2024-11-24T19:45:10,975 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/mapreduce.cluster.local.dir in system properties and HBase conf 2024-11-24T19:45:10,976 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-11-24T19:45:10,976 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-11-24T19:45:11,086 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-11-24T19:45:11,205 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-11-24T19:45:11,211 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-11-24T19:45:11,212 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-11-24T19:45:11,212 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-11-24T19:45:11,213 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-24T19:45:11,214 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-11-24T19:45:11,215 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-11-24T19:45:11,217 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-24T19:45:11,218 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-24T19:45:11,219 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-11-24T19:45:11,220 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/nfs.dump.dir in system properties and HBase conf 2024-11-24T19:45:11,220 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/java.io.tmpdir in system properties and HBase conf 2024-11-24T19:45:11,221 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-24T19:45:11,222 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-11-24T19:45:11,224 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-11-24T19:45:12,327 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-11-24T19:45:12,422 INFO [Time-limited test {}] log.Log(170): Logging initialized @2880ms to org.eclipse.jetty.util.log.Slf4jLog 2024-11-24T19:45:12,537 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:12,614 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-24T19:45:12,646 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-24T19:45:12,646 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-24T19:45:12,648 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-24T19:45:12,662 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:12,666 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@b03c34d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir/,AVAILABLE} 2024-11-24T19:45:12,667 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3cbd6fd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-24T19:45:12,886 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5cb83937{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/java.io.tmpdir/jetty-localhost-38387-hadoop-hdfs-3_4_1-tests_jar-_-any-14614029803549083038/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-24T19:45:12,899 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@69b5b273{HTTP/1.1, (http/1.1)}{localhost:38387} 2024-11-24T19:45:12,899 INFO [Time-limited test {}] server.Server(415): Started @3357ms 2024-11-24T19:45:13,576 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:13,586 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-24T19:45:13,591 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-24T19:45:13,591 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-24T19:45:13,591 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-24T19:45:13,594 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@179ed6d0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir/,AVAILABLE} 2024-11-24T19:45:13,595 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4f02cc61{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-24T19:45:13,708 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3990ff75{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/java.io.tmpdir/jetty-localhost-36963-hadoop-hdfs-3_4_1-tests_jar-_-any-11146629244314835880/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:13,709 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@48731e1b{HTTP/1.1, (http/1.1)}{localhost:36963} 2024-11-24T19:45:13,709 INFO [Time-limited test {}] server.Server(415): Started @4168ms 2024-11-24T19:45:13,775 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-24T19:45:13,912 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:13,919 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-24T19:45:13,927 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-24T19:45:13,928 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-24T19:45:13,928 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-24T19:45:13,930 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@50510811{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir/,AVAILABLE} 2024-11-24T19:45:13,931 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@125705fb{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-24T19:45:14,046 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@133f1bad{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/java.io.tmpdir/jetty-localhost-33921-hadoop-hdfs-3_4_1-tests_jar-_-any-12763881832705818383/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:14,047 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@39860596{HTTP/1.1, (http/1.1)}{localhost:33921} 2024-11-24T19:45:14,047 INFO [Time-limited test {}] server.Server(415): Started @4505ms 2024-11-24T19:45:14,051 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-24T19:45:14,099 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:14,105 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-24T19:45:14,107 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-24T19:45:14,108 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-24T19:45:14,108 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-24T19:45:14,111 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@28d0ee11{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir/,AVAILABLE} 2024-11-24T19:45:14,112 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@14c1b227{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-24T19:45:14,252 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@17f8e572{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/java.io.tmpdir/jetty-localhost-36907-hadoop-hdfs-3_4_1-tests_jar-_-any-5989980142238446703/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:14,253 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11aaae40{HTTP/1.1, (http/1.1)}{localhost:36907} 2024-11-24T19:45:14,254 INFO [Time-limited test {}] server.Server(415): Started @4712ms 2024-11-24T19:45:14,257 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-24T19:45:15,120 WARN [Thread-121 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data2/current/BP-29170651-172.17.0.2-1732477511872/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:15,120 WARN [Thread-120 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data1/current/BP-29170651-172.17.0.2-1732477511872/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:15,147 WARN [Thread-126 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data3/current/BP-29170651-172.17.0.2-1732477511872/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:15,147 WARN [Thread-127 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data4/current/BP-29170651-172.17.0.2-1732477511872/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:15,168 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-24T19:45:15,168 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-24T19:45:15,210 WARN [Thread-142 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data5/current/BP-29170651-172.17.0.2-1732477511872/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:15,211 WARN [Thread-143 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data6/current/BP-29170651-172.17.0.2-1732477511872/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:15,216 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x699b9678cf5a693a with lease ID 0x10105b071ce5b464: Processing first storage report for DS-bac2488f-7a87-40d7-a880-bb841fe627a5 from datanode DatanodeRegistration(127.0.0.1:37841, datanodeUuid=0277df7e-84cc-4e77-988c-6586260f6c5b, infoPort=41207, infoSecurePort=0, ipcPort=37701, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872) 2024-11-24T19:45:15,217 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x699b9678cf5a693a with lease ID 0x10105b071ce5b464: from storage DS-bac2488f-7a87-40d7-a880-bb841fe627a5 node DatanodeRegistration(127.0.0.1:37841, datanodeUuid=0277df7e-84cc-4e77-988c-6586260f6c5b, infoPort=41207, infoSecurePort=0, ipcPort=37701, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-24T19:45:15,218 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xbe9fcd2e568be7ce with lease ID 0x10105b071ce5b465: Processing first storage report for DS-31fcc325-ddbe-4d1e-875a-bee032695bae from datanode DatanodeRegistration(127.0.0.1:43851, datanodeUuid=665164a6-ecc0-44fd-8c67-f432dcd8b1f3, infoPort=40497, infoSecurePort=0, ipcPort=45941, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872) 2024-11-24T19:45:15,218 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbe9fcd2e568be7ce with lease ID 0x10105b071ce5b465: from storage DS-31fcc325-ddbe-4d1e-875a-bee032695bae node DatanodeRegistration(127.0.0.1:43851, datanodeUuid=665164a6-ecc0-44fd-8c67-f432dcd8b1f3, infoPort=40497, infoSecurePort=0, ipcPort=45941, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:15,218 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x699b9678cf5a693a with lease ID 0x10105b071ce5b464: Processing first storage report for DS-cd16a724-ded2-4e46-b397-8711fcd82c95 from datanode DatanodeRegistration(127.0.0.1:37841, datanodeUuid=0277df7e-84cc-4e77-988c-6586260f6c5b, infoPort=41207, infoSecurePort=0, ipcPort=37701, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872) 2024-11-24T19:45:15,219 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x699b9678cf5a693a with lease ID 0x10105b071ce5b464: from storage DS-cd16a724-ded2-4e46-b397-8711fcd82c95 node DatanodeRegistration(127.0.0.1:37841, datanodeUuid=0277df7e-84cc-4e77-988c-6586260f6c5b, infoPort=41207, infoSecurePort=0, ipcPort=37701, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:15,219 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xbe9fcd2e568be7ce with lease ID 0x10105b071ce5b465: Processing first storage report for DS-6d75a61d-d0d2-4f67-a75c-585db80ac703 from datanode DatanodeRegistration(127.0.0.1:43851, datanodeUuid=665164a6-ecc0-44fd-8c67-f432dcd8b1f3, infoPort=40497, infoSecurePort=0, ipcPort=45941, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872) 2024-11-24T19:45:15,219 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbe9fcd2e568be7ce with lease ID 0x10105b071ce5b465: from storage DS-6d75a61d-d0d2-4f67-a75c-585db80ac703 node DatanodeRegistration(127.0.0.1:43851, datanodeUuid=665164a6-ecc0-44fd-8c67-f432dcd8b1f3, infoPort=40497, infoSecurePort=0, ipcPort=45941, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:15,234 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-24T19:45:15,238 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7b7cbf8566ad7787 with lease ID 0x10105b071ce5b466: Processing first storage report for DS-6b7f112f-ebaf-495b-81e4-ae4c83840978 from datanode DatanodeRegistration(127.0.0.1:35937, datanodeUuid=2a332ccd-6d46-413a-8a7e-6880197ec257, infoPort=44049, infoSecurePort=0, ipcPort=34605, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872) 2024-11-24T19:45:15,239 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7b7cbf8566ad7787 with lease ID 0x10105b071ce5b466: from storage DS-6b7f112f-ebaf-495b-81e4-ae4c83840978 node DatanodeRegistration(127.0.0.1:35937, datanodeUuid=2a332ccd-6d46-413a-8a7e-6880197ec257, infoPort=44049, infoSecurePort=0, ipcPort=34605, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-24T19:45:15,239 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7b7cbf8566ad7787 with lease ID 0x10105b071ce5b466: Processing first storage report for DS-62428fa6-e863-4331-a712-7d1e4a5d82bb from datanode DatanodeRegistration(127.0.0.1:35937, datanodeUuid=2a332ccd-6d46-413a-8a7e-6880197ec257, infoPort=44049, infoSecurePort=0, ipcPort=34605, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872) 2024-11-24T19:45:15,239 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7b7cbf8566ad7787 with lease ID 0x10105b071ce5b466: from storage DS-62428fa6-e863-4331-a712-7d1e4a5d82bb node DatanodeRegistration(127.0.0.1:35937, datanodeUuid=2a332ccd-6d46-413a-8a7e-6880197ec257, infoPort=44049, infoSecurePort=0, ipcPort=34605, storageInfo=lv=-57;cid=testClusterID;nsid=1107505838;c=1732477511872), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:15,349 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6 2024-11-24T19:45:15,422 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-11-24T19:45:15,489 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=156, OpenFileDescriptor=393, MaxFileDescriptor=1048576, SystemLoadAverage=291, ProcessCount=11, AvailableMemoryMB=11172 2024-11-24T19:45:15,491 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-11-24T19:45:15,498 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-11-24T19:45:15,611 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/zookeeper_0, clientPort=61658, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-11-24T19:45:15,624 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=61658 2024-11-24T19:45:15,642 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:15,646 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:15,745 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:15,745 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:15,801 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:36702 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36702 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:15,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775792_1002 (size=7) 2024-11-24T19:45:16,222 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:16,233 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9 with version=8 2024-11-24T19:45:16,233 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/hbase-staging 2024-11-24T19:45:16,342 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-11-24T19:45:16,606 INFO [Time-limited test {}] client.ConnectionUtils(128): master/32a03c886245:0 server-side Connection retries=45 2024-11-24T19:45:16,617 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:16,617 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:16,624 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-24T19:45:16,624 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:16,624 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-24T19:45:16,799 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-11-24T19:45:16,863 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-11-24T19:45:16,874 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-11-24T19:45:16,879 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-24T19:45:16,909 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 18054 (auto-detected) 2024-11-24T19:45:16,910 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-11-24T19:45:16,933 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:33203 2024-11-24T19:45:16,954 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:33203 connecting to ZooKeeper ensemble=127.0.0.1:61658 2024-11-24T19:45:17,079 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:332030x0, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-24T19:45:17,081 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:33203-0x1016faa8a030000 connected 2024-11-24T19:45:17,166 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,169 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,177 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:17,183 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9, hbase.cluster.distributed=false 2024-11-24T19:45:17,214 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-24T19:45:17,220 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=33203 2024-11-24T19:45:17,220 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=33203 2024-11-24T19:45:17,221 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=33203 2024-11-24T19:45:17,221 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=33203 2024-11-24T19:45:17,221 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=33203 2024-11-24T19:45:17,328 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32a03c886245:0 server-side Connection retries=45 2024-11-24T19:45:17,331 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,331 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,331 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-24T19:45:17,332 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,332 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-24T19:45:17,335 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-24T19:45:17,337 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-24T19:45:17,339 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39441 2024-11-24T19:45:17,341 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39441 connecting to ZooKeeper ensemble=127.0.0.1:61658 2024-11-24T19:45:17,342 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,347 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,374 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:394410x0, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-24T19:45:17,375 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39441-0x1016faa8a030001 connected 2024-11-24T19:45:17,376 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:17,380 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-24T19:45:17,389 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-24T19:45:17,392 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-24T19:45:17,398 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-24T19:45:17,400 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39441 2024-11-24T19:45:17,400 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39441 2024-11-24T19:45:17,400 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39441 2024-11-24T19:45:17,401 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39441 2024-11-24T19:45:17,401 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39441 2024-11-24T19:45:17,418 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32a03c886245:0 server-side Connection retries=45 2024-11-24T19:45:17,418 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,418 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,419 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-24T19:45:17,419 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,419 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-24T19:45:17,420 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-24T19:45:17,420 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-24T19:45:17,421 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:38025 2024-11-24T19:45:17,424 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:38025 connecting to ZooKeeper ensemble=127.0.0.1:61658 2024-11-24T19:45:17,425 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,430 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,463 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:380250x0, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-24T19:45:17,463 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:38025-0x1016faa8a030002 connected 2024-11-24T19:45:17,464 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:17,464 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-24T19:45:17,465 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-24T19:45:17,466 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-24T19:45:17,469 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-24T19:45:17,471 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38025 2024-11-24T19:45:17,471 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38025 2024-11-24T19:45:17,472 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38025 2024-11-24T19:45:17,473 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38025 2024-11-24T19:45:17,473 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38025 2024-11-24T19:45:17,490 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32a03c886245:0 server-side Connection retries=45 2024-11-24T19:45:17,490 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,491 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,491 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-24T19:45:17,491 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:17,491 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-24T19:45:17,492 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-24T19:45:17,492 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-24T19:45:17,494 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:37807 2024-11-24T19:45:17,496 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:37807 connecting to ZooKeeper ensemble=127.0.0.1:61658 2024-11-24T19:45:17,497 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,499 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,513 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:378070x0, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-24T19:45:17,514 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37807-0x1016faa8a030003 connected 2024-11-24T19:45:17,515 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:17,515 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-24T19:45:17,519 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-24T19:45:17,520 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-24T19:45:17,522 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-24T19:45:17,523 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37807 2024-11-24T19:45:17,525 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37807 2024-11-24T19:45:17,527 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37807 2024-11-24T19:45:17,527 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37807 2024-11-24T19:45:17,528 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37807 2024-11-24T19:45:17,543 DEBUG [M:0;32a03c886245:33203 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;32a03c886245:33203 2024-11-24T19:45:17,544 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/32a03c886245,33203,1732477516430 2024-11-24T19:45:17,557 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:17,557 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:17,557 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:17,557 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:17,559 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/32a03c886245,33203,1732477516430 2024-11-24T19:45:17,587 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-24T19:45:17,587 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-24T19:45:17,588 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:17,588 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:17,588 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-24T19:45:17,588 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:17,588 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:17,589 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-11-24T19:45:17,591 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/32a03c886245,33203,1732477516430 from backup master directory 2024-11-24T19:45:17,598 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:17,598 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/32a03c886245,33203,1732477516430 2024-11-24T19:45:17,598 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:17,598 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:17,598 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:17,599 WARN [master/32a03c886245:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-24T19:45:17,599 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=32a03c886245,33203,1732477516430 2024-11-24T19:45:17,602 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-11-24T19:45:17,603 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-11-24T19:45:17,662 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/hbase.id] with ID: a3f7300c-71d6-4c6f-aa5b-9796d230d3e3 2024-11-24T19:45:17,662 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/.tmp/hbase.id 2024-11-24T19:45:17,669 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:17,669 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:17,674 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:36730 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36730 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:17,680 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775776_1004 (size=42) 2024-11-24T19:45:17,681 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:17,681 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/.tmp/hbase.id]:[hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/hbase.id] 2024-11-24T19:45:17,728 INFO [master/32a03c886245:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:17,733 INFO [master/32a03c886245:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-11-24T19:45:17,751 INFO [master/32a03c886245:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 16ms. 2024-11-24T19:45:17,773 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:17,773 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:17,773 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:17,773 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:17,790 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:17,790 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:17,795 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:35534 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:37841:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35534 dst: /127.0.0.1:37841 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:17,800 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_-9223372036854775760_1006 (size=196) 2024-11-24T19:45:17,801 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:17,817 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-24T19:45:17,819 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-11-24T19:45:17,824 INFO [master/32a03c886245:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-24T19:45:17,857 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:17,858 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:17,865 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:36754 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36754 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:17,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775744_1008 (size=1189) 2024-11-24T19:45:17,871 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:17,887 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store 2024-11-24T19:45:17,905 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:17,906 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:17,909 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:35136 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:35937:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35136 dst: /127.0.0.1:35937 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:17,915 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775728_1010 (size=34) 2024-11-24T19:45:18,263 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775789_1002 (size=7) 2024-11-24T19:45:18,263 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_-9223372036854775788_1002 (size=7) 2024-11-24T19:45:18,317 WARN [master/32a03c886245:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:18,321 INFO [master/32a03c886245:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-11-24T19:45:18,324 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:18,326 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-24T19:45:18,326 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:18,327 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:18,329 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-24T19:45:18,330 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:18,330 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:18,331 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1732477518326Disabling compacts and flushes for region at 1732477518326Disabling writes for close at 1732477518329 (+3 ms)Writing region close event to WAL at 1732477518330 (+1 ms)Closed at 1732477518330 2024-11-24T19:45:18,334 WARN [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/.initializing 2024-11-24T19:45:18,334 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/WALs/32a03c886245,33203,1732477516430 2024-11-24T19:45:18,345 INFO [master/32a03c886245:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-24T19:45:18,363 INFO [master/32a03c886245:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C33203%2C1732477516430, suffix=, logDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/WALs/32a03c886245,33203,1732477516430, archiveDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/oldWALs, maxLogs=10 2024-11-24T19:45:18,405 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/WALs/32a03c886245,33203,1732477516430/32a03c886245%2C33203%2C1732477516430.1732477518370, exclude list is [], retry=0 2024-11-24T19:45:18,410 WARN [IPC Server handler 2 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:18,410 WARN [IPC Server handler 2 on default port 40555 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:18,410 WARN [IPC Server handler 2 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:18,428 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:18,429 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35937,DS-6b7f112f-ebaf-495b-81e4-ae4c83840978,DISK] 2024-11-24T19:45:18,429 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37841,DS-bac2488f-7a87-40d7-a880-bb841fe627a5,DISK] 2024-11-24T19:45:18,433 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-11-24T19:45:18,477 INFO [master/32a03c886245:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/WALs/32a03c886245,33203,1732477516430/32a03c886245%2C33203%2C1732477516430.1732477518370 2024-11-24T19:45:18,478 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:44049:44049),(127.0.0.1/127.0.0.1:41207:41207)] 2024-11-24T19:45:18,479 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-11-24T19:45:18,479 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:18,483 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,484 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,525 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,552 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-11-24T19:45:18,556 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:18,559 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:18,559 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,564 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-11-24T19:45:18,564 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:18,565 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-24T19:45:18,565 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,569 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-11-24T19:45:18,569 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:18,570 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-24T19:45:18,571 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,575 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-11-24T19:45:18,576 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:18,577 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-24T19:45:18,578 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,581 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,583 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,588 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,589 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,593 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-24T19:45:18,597 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:18,614 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-24T19:45:18,616 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70659577, jitterRate=0.05290974676609039}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-24T19:45:18,626 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1732477518500Initializing all the Stores at 1732477518502 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477518503 (+1 ms)Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477518503Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477518504 (+1 ms)Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477518504Cleaning up temporary data from old regions at 1732477518589 (+85 ms)Region opened successfully at 1732477518625 (+36 ms) 2024-11-24T19:45:18,628 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-11-24T19:45:18,676 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@265f495b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32a03c886245/172.17.0.2:0 2024-11-24T19:45:18,714 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-11-24T19:45:18,726 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-11-24T19:45:18,727 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-11-24T19:45:18,730 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-11-24T19:45:18,731 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-11-24T19:45:18,739 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 6 msec 2024-11-24T19:45:18,739 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-11-24T19:45:18,772 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-11-24T19:45:18,783 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-11-24T19:45:18,845 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-11-24T19:45:18,848 INFO [master/32a03c886245:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-11-24T19:45:18,849 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-11-24T19:45:18,856 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-11-24T19:45:18,859 INFO [master/32a03c886245:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-11-24T19:45:18,862 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-11-24T19:45:18,873 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-11-24T19:45:18,875 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-11-24T19:45:18,881 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-11-24T19:45:18,903 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-11-24T19:45:18,915 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-11-24T19:45:18,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:18,929 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:18,929 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:18,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:18,929 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:18,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:18,929 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:18,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:18,933 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=32a03c886245,33203,1732477516430, sessionid=0x1016faa8a030000, setting cluster-up flag (Was=false) 2024-11-24T19:45:18,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:18,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:18,956 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:18,956 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:18,981 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-11-24T19:45:18,984 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=32a03c886245,33203,1732477516430 2024-11-24T19:45:19,006 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:19,006 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:19,006 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:19,006 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:19,031 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-11-24T19:45:19,033 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=32a03c886245,33203,1732477516430 2024-11-24T19:45:19,039 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-11-24T19:45:19,126 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-11-24T19:45:19,138 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(746): ClusterId : a3f7300c-71d6-4c6f-aa5b-9796d230d3e3 2024-11-24T19:45:19,138 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(746): ClusterId : a3f7300c-71d6-4c6f-aa5b-9796d230d3e3 2024-11-24T19:45:19,138 INFO [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(746): ClusterId : a3f7300c-71d6-4c6f-aa5b-9796d230d3e3 2024-11-24T19:45:19,138 INFO [master/32a03c886245:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-11-24T19:45:19,141 DEBUG [RS:2;32a03c886245:37807 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-24T19:45:19,141 DEBUG [RS:0;32a03c886245:39441 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-24T19:45:19,141 DEBUG [RS:1;32a03c886245:38025 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-24T19:45:19,145 INFO [master/32a03c886245:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-11-24T19:45:19,152 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 32a03c886245,33203,1732477516430 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-11-24T19:45:19,164 DEBUG [RS:1;32a03c886245:38025 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-24T19:45:19,164 DEBUG [RS:2;32a03c886245:37807 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-24T19:45:19,164 DEBUG [RS:0;32a03c886245:39441 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-24T19:45:19,164 DEBUG [RS:1;32a03c886245:38025 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-24T19:45:19,164 DEBUG [RS:2;32a03c886245:37807 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-24T19:45:19,164 DEBUG [RS:0;32a03c886245:39441 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-24T19:45:19,166 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/32a03c886245:0, corePoolSize=5, maxPoolSize=5 2024-11-24T19:45:19,166 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/32a03c886245:0, corePoolSize=5, maxPoolSize=5 2024-11-24T19:45:19,166 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/32a03c886245:0, corePoolSize=5, maxPoolSize=5 2024-11-24T19:45:19,166 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/32a03c886245:0, corePoolSize=5, maxPoolSize=5 2024-11-24T19:45:19,167 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/32a03c886245:0, corePoolSize=10, maxPoolSize=10 2024-11-24T19:45:19,167 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,167 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/32a03c886245:0, corePoolSize=2, maxPoolSize=2 2024-11-24T19:45:19,167 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,171 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1732477549171 2024-11-24T19:45:19,174 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-11-24T19:45:19,174 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-24T19:45:19,175 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-11-24T19:45:19,175 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-11-24T19:45:19,179 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-11-24T19:45:19,180 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-11-24T19:45:19,180 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-11-24T19:45:19,180 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-11-24T19:45:19,181 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:19,181 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-11-24T19:45:19,182 DEBUG [RS:1;32a03c886245:38025 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-24T19:45:19,183 DEBUG [RS:2;32a03c886245:37807 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-24T19:45:19,183 DEBUG [RS:1;32a03c886245:38025 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2c5a647c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32a03c886245/172.17.0.2:0 2024-11-24T19:45:19,183 DEBUG [RS:2;32a03c886245:37807 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@ae41bdb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32a03c886245/172.17.0.2:0 2024-11-24T19:45:19,186 DEBUG [RS:0;32a03c886245:39441 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-24T19:45:19,186 DEBUG [RS:0;32a03c886245:39441 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@15b3fc4c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32a03c886245/172.17.0.2:0 2024-11-24T19:45:19,186 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,201 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-11-24T19:45:19,202 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-11-24T19:45:19,203 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-11-24T19:45:19,207 DEBUG [RS:2;32a03c886245:37807 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;32a03c886245:37807 2024-11-24T19:45:19,210 DEBUG [RS:1;32a03c886245:38025 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;32a03c886245:38025 2024-11-24T19:45:19,210 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-11-24T19:45:19,210 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-11-24T19:45:19,211 INFO [RS:2;32a03c886245:37807 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-24T19:45:19,211 INFO [RS:1;32a03c886245:38025 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-24T19:45:19,211 INFO [RS:2;32a03c886245:37807 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-24T19:45:19,211 INFO [RS:1;32a03c886245:38025 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-24T19:45:19,211 DEBUG [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-24T19:45:19,211 DEBUG [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-24T19:45:19,214 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(2659): reportForDuty to master=32a03c886245,33203,1732477516430 with port=37807, startcode=1732477517489 2024-11-24T19:45:19,214 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(2659): reportForDuty to master=32a03c886245,33203,1732477516430 with port=38025, startcode=1732477517418 2024-11-24T19:45:19,215 DEBUG [RS:0;32a03c886245:39441 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;32a03c886245:39441 2024-11-24T19:45:19,216 INFO [RS:0;32a03c886245:39441 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-24T19:45:19,216 INFO [RS:0;32a03c886245:39441 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-24T19:45:19,216 DEBUG [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-24T19:45:19,218 INFO [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(2659): reportForDuty to master=32a03c886245,33203,1732477516430 with port=39441, startcode=1732477517291 2024-11-24T19:45:19,218 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:19,219 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:19,222 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/32a03c886245:0:becomeActiveMaster-HFileCleaner.large.0-1732477519212,5,FailOnTimeoutGroup] 2024-11-24T19:45:19,223 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/32a03c886245:0:becomeActiveMaster-HFileCleaner.small.0-1732477519223,5,FailOnTimeoutGroup] 2024-11-24T19:45:19,223 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,224 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-11-24T19:45:19,225 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,225 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,227 DEBUG [RS:1;32a03c886245:38025 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-24T19:45:19,227 DEBUG [RS:2;32a03c886245:37807 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-24T19:45:19,231 DEBUG [RS:0;32a03c886245:39441 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-24T19:45:19,266 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:36770 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36770 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:19,275 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775712_1013 (size=1321) 2024-11-24T19:45:19,277 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:47903, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-11-24T19:45:19,277 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50429, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-11-24T19:45:19,277 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:19,279 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:40587, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-11-24T19:45:19,279 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-11-24T19:45:19,280 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9 2024-11-24T19:45:19,286 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33203 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32a03c886245,37807,1732477517489 2024-11-24T19:45:19,289 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33203 {}] master.ServerManager(517): Registering regionserver=32a03c886245,37807,1732477517489 2024-11-24T19:45:19,297 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:19,297 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:19,303 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33203 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32a03c886245,38025,1732477517418 2024-11-24T19:45:19,303 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33203 {}] master.ServerManager(517): Registering regionserver=32a03c886245,38025,1732477517418 2024-11-24T19:45:19,307 DEBUG [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9 2024-11-24T19:45:19,307 DEBUG [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40555 2024-11-24T19:45:19,307 DEBUG [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-24T19:45:19,308 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33203 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32a03c886245,39441,1732477517291 2024-11-24T19:45:19,308 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=33203 {}] master.ServerManager(517): Registering regionserver=32a03c886245,39441,1732477517291 2024-11-24T19:45:19,308 DEBUG [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9 2024-11-24T19:45:19,308 DEBUG [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40555 2024-11-24T19:45:19,308 DEBUG [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-24T19:45:19,311 DEBUG [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9 2024-11-24T19:45:19,311 DEBUG [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40555 2024-11-24T19:45:19,312 DEBUG [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-24T19:45:19,318 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:36796 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36796 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:19,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775696_1015 (size=32) 2024-11-24T19:45:19,333 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:19,334 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:19,339 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-24T19:45:19,342 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-24T19:45:19,342 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:19,344 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:19,344 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-24T19:45:19,349 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-24T19:45:19,349 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:19,351 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:19,351 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-24T19:45:19,354 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-24T19:45:19,355 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:19,356 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:19,356 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-24T19:45:19,360 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-24T19:45:19,360 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:19,361 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-24T19:45:19,363 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:19,363 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-24T19:45:19,366 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740 2024-11-24T19:45:19,367 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740 2024-11-24T19:45:19,370 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-24T19:45:19,371 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-24T19:45:19,372 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-24T19:45:19,375 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-24T19:45:19,381 DEBUG [RS:2;32a03c886245:37807 {}] zookeeper.ZKUtil(111): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32a03c886245,37807,1732477517489 2024-11-24T19:45:19,381 DEBUG [RS:1;32a03c886245:38025 {}] zookeeper.ZKUtil(111): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32a03c886245,38025,1732477517418 2024-11-24T19:45:19,382 WARN [RS:2;32a03c886245:37807 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-24T19:45:19,382 WARN [RS:1;32a03c886245:38025 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-24T19:45:19,382 INFO [RS:2;32a03c886245:37807 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-24T19:45:19,382 INFO [RS:1;32a03c886245:38025 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-24T19:45:19,382 DEBUG [RS:0;32a03c886245:39441 {}] zookeeper.ZKUtil(111): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32a03c886245,39441,1732477517291 2024-11-24T19:45:19,382 WARN [RS:0;32a03c886245:39441 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-24T19:45:19,382 INFO [RS:0;32a03c886245:39441 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-24T19:45:19,382 DEBUG [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,38025,1732477517418 2024-11-24T19:45:19,382 DEBUG [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,37807,1732477517489 2024-11-24T19:45:19,382 DEBUG [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,39441,1732477517291 2024-11-24T19:45:19,384 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32a03c886245,39441,1732477517291] 2024-11-24T19:45:19,384 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32a03c886245,37807,1732477517489] 2024-11-24T19:45:19,384 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32a03c886245,38025,1732477517418] 2024-11-24T19:45:19,414 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-24T19:45:19,415 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61331833, jitterRate=-0.08608447015285492}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-24T19:45:19,420 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1732477519334Initializing all the Stores at 1732477519336 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477519336Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477519339 (+3 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477519339Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477519339Cleaning up temporary data from old regions at 1732477519371 (+32 ms)Region opened successfully at 1732477519420 (+49 ms) 2024-11-24T19:45:19,421 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-24T19:45:19,421 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-24T19:45:19,421 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-24T19:45:19,421 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-24T19:45:19,421 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-24T19:45:19,425 INFO [RS:2;32a03c886245:37807 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-24T19:45:19,425 INFO [RS:0;32a03c886245:39441 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-24T19:45:19,425 INFO [RS:1;32a03c886245:38025 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-24T19:45:19,427 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-24T19:45:19,428 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1732477519421Disabling compacts and flushes for region at 1732477519421Disabling writes for close at 1732477519421Writing region close event to WAL at 1732477519427 (+6 ms)Closed at 1732477519427 2024-11-24T19:45:19,431 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-24T19:45:19,431 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-11-24T19:45:19,438 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-11-24T19:45:19,449 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-24T19:45:19,451 INFO [RS:2;32a03c886245:37807 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-24T19:45:19,451 INFO [RS:0;32a03c886245:39441 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-24T19:45:19,453 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-11-24T19:45:19,455 INFO [RS:1;32a03c886245:38025 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-24T19:45:19,458 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=2) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-24T19:45:19,458 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:19,458 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=1) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-24T19:45:19,459 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:19,459 WARN [RedundancyMonitor {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK, ARCHIVE], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:19,459 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) All required storage types are unavailable: unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:19,459 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=2) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-24T19:45:19,459 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:19,460 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=1) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-24T19:45:19,460 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:19,460 WARN [RedundancyMonitor {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK, ARCHIVE], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:19,460 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) All required storage types are unavailable: unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:19,466 INFO [RS:2;32a03c886245:37807 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-24T19:45:19,466 INFO [RS:0;32a03c886245:39441 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-24T19:45:19,466 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,466 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,467 INFO [RS:1;32a03c886245:38025 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-24T19:45:19,467 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,468 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-24T19:45:19,476 INFO [RS:1;32a03c886245:38025 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-24T19:45:19,477 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-24T19:45:19,478 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,479 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,479 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,479 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,479 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,479 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,480 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32a03c886245:0, corePoolSize=2, maxPoolSize=2 2024-11-24T19:45:19,480 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,480 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,480 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,480 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,480 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,480 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,481 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:19,481 DEBUG [RS:1;32a03c886245:38025 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:19,482 INFO [RS:2;32a03c886245:37807 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-24T19:45:19,482 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,483 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,483 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,483 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,483 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,483 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,483 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32a03c886245:0, corePoolSize=2, maxPoolSize=2 2024-11-24T19:45:19,483 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,483 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,484 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,484 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,484 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,484 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,484 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:19,484 DEBUG [RS:2;32a03c886245:37807 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:19,486 INFO [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-24T19:45:19,489 INFO [RS:0;32a03c886245:39441 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-24T19:45:19,489 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,489 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,489 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,489 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,489 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,490 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,490 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32a03c886245:0, corePoolSize=2, maxPoolSize=2 2024-11-24T19:45:19,490 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,490 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,490 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,490 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,490 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,491 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:19,491 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:19,491 DEBUG [RS:0;32a03c886245:39441 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:19,503 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,503 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,504 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,504 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,504 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,504 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,39441,1732477517291-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-24T19:45:19,508 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,508 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,508 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,509 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,509 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,509 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,37807,1732477517489-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-24T19:45:19,510 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,511 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,511 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,511 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,511 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,511 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,38025,1732477517418-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-24T19:45:19,542 INFO [RS:0;32a03c886245:39441 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-24T19:45:19,542 INFO [RS:2;32a03c886245:37807 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-24T19:45:19,545 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,37807,1732477517489-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,545 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,39441,1732477517291-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,545 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,545 INFO [RS:2;32a03c886245:37807 {}] regionserver.Replication(171): 32a03c886245,37807,1732477517489 started 2024-11-24T19:45:19,545 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,546 INFO [RS:0;32a03c886245:39441 {}] regionserver.Replication(171): 32a03c886245,39441,1732477517291 started 2024-11-24T19:45:19,546 INFO [RS:1;32a03c886245:38025 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-24T19:45:19,546 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,38025,1732477517418-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,546 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,546 INFO [RS:1;32a03c886245:38025 {}] regionserver.Replication(171): 32a03c886245,38025,1732477517418 started 2024-11-24T19:45:19,572 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,572 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,572 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1482): Serving as 32a03c886245,38025,1732477517418, RpcServer on 32a03c886245/172.17.0.2:38025, sessionid=0x1016faa8a030002 2024-11-24T19:45:19,572 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1482): Serving as 32a03c886245,37807,1732477517489, RpcServer on 32a03c886245/172.17.0.2:37807, sessionid=0x1016faa8a030003 2024-11-24T19:45:19,573 DEBUG [RS:2;32a03c886245:37807 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-24T19:45:19,573 DEBUG [RS:1;32a03c886245:38025 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-24T19:45:19,573 DEBUG [RS:1;32a03c886245:38025 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32a03c886245,38025,1732477517418 2024-11-24T19:45:19,573 DEBUG [RS:2;32a03c886245:37807 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32a03c886245,37807,1732477517489 2024-11-24T19:45:19,574 DEBUG [RS:1;32a03c886245:38025 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,38025,1732477517418' 2024-11-24T19:45:19,574 DEBUG [RS:2;32a03c886245:37807 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,37807,1732477517489' 2024-11-24T19:45:19,574 DEBUG [RS:2;32a03c886245:37807 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-24T19:45:19,574 DEBUG [RS:1;32a03c886245:38025 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-24T19:45:19,574 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:19,574 INFO [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(1482): Serving as 32a03c886245,39441,1732477517291, RpcServer on 32a03c886245/172.17.0.2:39441, sessionid=0x1016faa8a030001 2024-11-24T19:45:19,575 DEBUG [RS:0;32a03c886245:39441 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-24T19:45:19,575 DEBUG [RS:0;32a03c886245:39441 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32a03c886245,39441,1732477517291 2024-11-24T19:45:19,575 DEBUG [RS:0;32a03c886245:39441 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,39441,1732477517291' 2024-11-24T19:45:19,575 DEBUG [RS:0;32a03c886245:39441 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-24T19:45:19,575 DEBUG [RS:1;32a03c886245:38025 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-24T19:45:19,576 DEBUG [RS:0;32a03c886245:39441 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-24T19:45:19,575 DEBUG [RS:2;32a03c886245:37807 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-24T19:45:19,576 DEBUG [RS:1;32a03c886245:38025 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-24T19:45:19,577 DEBUG [RS:1;32a03c886245:38025 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-24T19:45:19,578 DEBUG [RS:1;32a03c886245:38025 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32a03c886245,38025,1732477517418 2024-11-24T19:45:19,578 DEBUG [RS:1;32a03c886245:38025 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,38025,1732477517418' 2024-11-24T19:45:19,578 DEBUG [RS:1;32a03c886245:38025 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-24T19:45:19,578 DEBUG [RS:0;32a03c886245:39441 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-24T19:45:19,578 DEBUG [RS:0;32a03c886245:39441 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-24T19:45:19,579 DEBUG [RS:0;32a03c886245:39441 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32a03c886245,39441,1732477517291 2024-11-24T19:45:19,579 DEBUG [RS:0;32a03c886245:39441 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,39441,1732477517291' 2024-11-24T19:45:19,579 DEBUG [RS:0;32a03c886245:39441 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-24T19:45:19,581 DEBUG [RS:1;32a03c886245:38025 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-24T19:45:19,581 DEBUG [RS:2;32a03c886245:37807 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-24T19:45:19,581 DEBUG [RS:2;32a03c886245:37807 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-24T19:45:19,582 DEBUG [RS:2;32a03c886245:37807 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32a03c886245,37807,1732477517489 2024-11-24T19:45:19,582 DEBUG [RS:2;32a03c886245:37807 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,37807,1732477517489' 2024-11-24T19:45:19,582 DEBUG [RS:2;32a03c886245:37807 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-24T19:45:19,582 DEBUG [RS:1;32a03c886245:38025 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-24T19:45:19,582 INFO [RS:1;32a03c886245:38025 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-24T19:45:19,582 INFO [RS:1;32a03c886245:38025 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-24T19:45:19,583 DEBUG [RS:2;32a03c886245:37807 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-24T19:45:19,583 DEBUG [RS:0;32a03c886245:39441 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-24T19:45:19,583 DEBUG [RS:2;32a03c886245:37807 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-24T19:45:19,583 INFO [RS:2;32a03c886245:37807 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-24T19:45:19,583 INFO [RS:2;32a03c886245:37807 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-24T19:45:19,584 DEBUG [RS:0;32a03c886245:39441 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-24T19:45:19,584 INFO [RS:0;32a03c886245:39441 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-24T19:45:19,584 INFO [RS:0;32a03c886245:39441 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-24T19:45:19,605 WARN [32a03c886245:33203 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-11-24T19:45:19,687 INFO [RS:1;32a03c886245:38025 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-24T19:45:19,687 INFO [RS:2;32a03c886245:37807 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-24T19:45:19,687 INFO [RS:0;32a03c886245:39441 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-24T19:45:19,690 INFO [RS:0;32a03c886245:39441 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C39441%2C1732477517291, suffix=, logDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,39441,1732477517291, archiveDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/oldWALs, maxLogs=32 2024-11-24T19:45:19,690 INFO [RS:2;32a03c886245:37807 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C37807%2C1732477517489, suffix=, logDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,37807,1732477517489, archiveDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/oldWALs, maxLogs=32 2024-11-24T19:45:19,690 INFO [RS:1;32a03c886245:38025 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C38025%2C1732477517418, suffix=, logDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,38025,1732477517418, archiveDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/oldWALs, maxLogs=32 2024-11-24T19:45:19,710 DEBUG [RS:2;32a03c886245:37807 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,37807,1732477517489/32a03c886245%2C37807%2C1732477517489.1732477519694, exclude list is [], retry=0 2024-11-24T19:45:19,716 DEBUG [RS:1;32a03c886245:38025 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,38025,1732477517418/32a03c886245%2C38025%2C1732477517418.1732477519694, exclude list is [], retry=0 2024-11-24T19:45:19,716 DEBUG [RS:0;32a03c886245:39441 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,39441,1732477517291/32a03c886245%2C39441%2C1732477517291.1732477519695, exclude list is [], retry=0 2024-11-24T19:45:19,719 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35937,DS-6b7f112f-ebaf-495b-81e4-ae4c83840978,DISK] 2024-11-24T19:45:19,719 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43851,DS-31fcc325-ddbe-4d1e-875a-bee032695bae,DISK] 2024-11-24T19:45:19,721 WARN [IPC Server handler 2 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:19,721 WARN [IPC Server handler 0 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:19,721 WARN [IPC Server handler 2 on default port 40555 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:19,721 WARN [IPC Server handler 0 on default port 40555 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:19,721 WARN [IPC Server handler 2 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:19,721 WARN [IPC Server handler 0 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:19,724 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37841,DS-bac2488f-7a87-40d7-a880-bb841fe627a5,DISK] 2024-11-24T19:45:19,725 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35937,DS-6b7f112f-ebaf-495b-81e4-ae4c83840978,DISK] 2024-11-24T19:45:19,726 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37841,DS-bac2488f-7a87-40d7-a880-bb841fe627a5,DISK] 2024-11-24T19:45:19,727 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37841,DS-bac2488f-7a87-40d7-a880-bb841fe627a5,DISK] 2024-11-24T19:45:19,728 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35937,DS-6b7f112f-ebaf-495b-81e4-ae4c83840978,DISK] 2024-11-24T19:45:19,761 INFO [RS:1;32a03c886245:38025 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,38025,1732477517418/32a03c886245%2C38025%2C1732477517418.1732477519694 2024-11-24T19:45:19,761 INFO [RS:2;32a03c886245:37807 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,37807,1732477517489/32a03c886245%2C37807%2C1732477517489.1732477519694 2024-11-24T19:45:19,761 INFO [RS:0;32a03c886245:39441 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,39441,1732477517291/32a03c886245%2C39441%2C1732477517291.1732477519695 2024-11-24T19:45:19,761 DEBUG [RS:1;32a03c886245:38025 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:44049:44049),(127.0.0.1/127.0.0.1:41207:41207)] 2024-11-24T19:45:19,762 DEBUG [RS:2;32a03c886245:37807 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:40497:40497),(127.0.0.1/127.0.0.1:41207:41207),(127.0.0.1/127.0.0.1:44049:44049)] 2024-11-24T19:45:19,765 DEBUG [RS:0;32a03c886245:39441 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:41207:41207),(127.0.0.1/127.0.0.1:44049:44049)] 2024-11-24T19:45:19,857 DEBUG [32a03c886245:33203 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-11-24T19:45:19,865 DEBUG [32a03c886245:33203 {}] balancer.BalancerClusterState(204): Hosts are {32a03c886245=0} racks are {/default-rack=0} 2024-11-24T19:45:19,871 DEBUG [32a03c886245:33203 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-24T19:45:19,871 DEBUG [32a03c886245:33203 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-24T19:45:19,871 DEBUG [32a03c886245:33203 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-24T19:45:19,871 DEBUG [32a03c886245:33203 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-24T19:45:19,871 DEBUG [32a03c886245:33203 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-24T19:45:19,871 DEBUG [32a03c886245:33203 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-24T19:45:19,871 INFO [32a03c886245:33203 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-24T19:45:19,871 INFO [32a03c886245:33203 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-24T19:45:19,871 INFO [32a03c886245:33203 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-24T19:45:19,871 DEBUG [32a03c886245:33203 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-24T19:45:19,878 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=32a03c886245,38025,1732477517418 2024-11-24T19:45:19,887 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 32a03c886245,38025,1732477517418, state=OPENING 2024-11-24T19:45:19,906 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-11-24T19:45:19,914 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:19,914 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:19,914 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:19,914 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:19,915 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:19,915 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:19,915 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:19,915 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:19,918 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-24T19:45:19,920 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=32a03c886245,38025,1732477517418}] 2024-11-24T19:45:20,095 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-24T19:45:20,098 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37087, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-24T19:45:20,110 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-11-24T19:45:20,110 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-24T19:45:20,111 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-11-24T19:45:20,116 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C38025%2C1732477517418.meta, suffix=.meta, logDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,38025,1732477517418, archiveDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/oldWALs, maxLogs=32 2024-11-24T19:45:20,131 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,38025,1732477517418/32a03c886245%2C38025%2C1732477517418.meta.1732477520117.meta, exclude list is [], retry=0 2024-11-24T19:45:20,134 WARN [IPC Server handler 0 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:20,134 WARN [IPC Server handler 0 on default port 40555 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:20,134 WARN [IPC Server handler 0 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:20,136 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37841,DS-bac2488f-7a87-40d7-a880-bb841fe627a5,DISK] 2024-11-24T19:45:20,136 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35937,DS-6b7f112f-ebaf-495b-81e4-ae4c83840978,DISK] 2024-11-24T19:45:20,143 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/WALs/32a03c886245,38025,1732477517418/32a03c886245%2C38025%2C1732477517418.meta.1732477520117.meta 2024-11-24T19:45:20,145 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:44049:44049),(127.0.0.1/127.0.0.1:41207:41207)] 2024-11-24T19:45:20,145 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-11-24T19:45:20,147 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-11-24T19:45:20,149 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-11-24T19:45:20,155 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-11-24T19:45:20,160 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-11-24T19:45:20,161 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:20,161 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-11-24T19:45:20,161 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-11-24T19:45:20,165 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-24T19:45:20,166 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-24T19:45:20,167 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:20,167 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:20,168 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-24T19:45:20,169 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-24T19:45:20,169 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:20,170 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:20,170 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-24T19:45:20,171 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-24T19:45:20,172 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:20,173 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:20,173 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-24T19:45:20,175 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-24T19:45:20,175 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:20,177 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:20,177 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-24T19:45:20,179 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740 2024-11-24T19:45:20,183 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740 2024-11-24T19:45:20,185 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-24T19:45:20,186 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-24T19:45:20,186 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-24T19:45:20,190 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-24T19:45:20,192 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65470199, jitterRate=-0.02441801130771637}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-24T19:45:20,192 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-11-24T19:45:20,194 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1732477520162Writing region info on filesystem at 1732477520162Initializing all the Stores at 1732477520164 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477520164Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477520164Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477520164Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477520165 (+1 ms)Cleaning up temporary data from old regions at 1732477520186 (+21 ms)Running coprocessor post-open hooks at 1732477520192 (+6 ms)Region opened successfully at 1732477520194 (+2 ms) 2024-11-24T19:45:20,203 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1732477520087 2024-11-24T19:45:20,216 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-11-24T19:45:20,217 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-11-24T19:45:20,219 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=32a03c886245,38025,1732477517418 2024-11-24T19:45:20,221 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 32a03c886245,38025,1732477517418, state=OPEN 2024-11-24T19:45:20,287 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-24T19:45:20,287 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-24T19:45:20,287 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-24T19:45:20,287 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-24T19:45:20,287 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:20,287 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:20,287 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:20,287 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:20,288 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=32a03c886245,38025,1732477517418 2024-11-24T19:45:20,293 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-11-24T19:45:20,293 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=32a03c886245,38025,1732477517418 in 368 msec 2024-11-24T19:45:20,306 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-11-24T19:45:20,306 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 858 msec 2024-11-24T19:45:20,308 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-24T19:45:20,309 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-11-24T19:45:20,329 DEBUG [PEWorker-1 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-24T19:45:20,330 DEBUG [PEWorker-1 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=32a03c886245,38025,1732477517418, seqNum=-1] 2024-11-24T19:45:20,380 DEBUG [PEWorker-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-24T19:45:20,383 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49559, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-24T19:45:20,415 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.3410 sec 2024-11-24T19:45:20,416 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1732477520416, completionTime=-1 2024-11-24T19:45:20,418 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-11-24T19:45:20,418 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-11-24T19:45:20,441 INFO [master/32a03c886245:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-11-24T19:45:20,441 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1732477580441 2024-11-24T19:45:20,441 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1732477640441 2024-11-24T19:45:20,441 INFO [master/32a03c886245:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 22 msec 2024-11-24T19:45:20,442 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-11-24T19:45:20,451 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,33203,1732477516430-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:20,451 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,33203,1732477516430-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:20,451 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,33203,1732477516430-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:20,453 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-32a03c886245:33203, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:20,453 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:20,454 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:20,460 DEBUG [master/32a03c886245:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-11-24T19:45:20,484 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.884sec 2024-11-24T19:45:20,486 INFO [master/32a03c886245:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-11-24T19:45:20,487 INFO [master/32a03c886245:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-11-24T19:45:20,488 INFO [master/32a03c886245:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-11-24T19:45:20,489 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-11-24T19:45:20,490 INFO [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-11-24T19:45:20,491 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,33203,1732477516430-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-24T19:45:20,492 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,33203,1732477516430-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-11-24T19:45:20,496 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-11-24T19:45:20,497 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-11-24T19:45:20,497 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,33203,1732477516430-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:20,568 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1ee231f4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-24T19:45:20,571 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-11-24T19:45:20,571 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-11-24T19:45:20,575 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 32a03c886245,33203,-1 for getting cluster id 2024-11-24T19:45:20,577 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-11-24T19:45:20,585 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'a3f7300c-71d6-4c6f-aa5b-9796d230d3e3' 2024-11-24T19:45:20,588 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-11-24T19:45:20,588 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "a3f7300c-71d6-4c6f-aa5b-9796d230d3e3" 2024-11-24T19:45:20,589 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@18858350, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-24T19:45:20,589 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [32a03c886245,33203,-1] 2024-11-24T19:45:20,592 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-11-24T19:45:20,594 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:20,595 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43236, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-11-24T19:45:20,598 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@442c1f0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-24T19:45:20,598 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-24T19:45:20,606 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=32a03c886245,38025,1732477517418, seqNum=-1] 2024-11-24T19:45:20,607 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-24T19:45:20,612 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59254, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-24T19:45:20,635 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=32a03c886245,33203,1732477516430 2024-11-24T19:45:20,638 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-11-24T19:45:20,643 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 32a03c886245,33203,1732477516430 2024-11-24T19:45:20,646 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@20e384d9 2024-11-24T19:45:20,647 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-11-24T19:45:20,650 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43246, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-11-24T19:45:20,657 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-24T19:45:20,666 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-11-24T19:45:20,669 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-11-24T19:45:20,671 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-11-24T19:45:20,672 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:20,674 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-24T19:45:20,677 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:20,685 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:20,685 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:20,692 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:36818 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36818 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:20,698 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775680_1021 (size=392) 2024-11-24T19:45:20,698 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:20,701 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 005eb91681f1bb20289d5dd15edd33d3, NAME => 'TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9 2024-11-24T19:45:20,710 WARN [IPC Server handler 2 on default port 40555 {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=3) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-24T19:45:20,710 WARN [IPC Server handler 2 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:20,710 WARN [IPC Server handler 2 on default port 40555 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:20,710 WARN [IPC Server handler 2 on default port 40555 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:20,714 WARN [PEWorker-3 {}] procedure.CreateTableProcedure(152): Retriable error trying to create table=TestHBaseWalOnEC,state=CREATE_TABLE_WRITE_FS_LAYOUT,suspend 1secs. java.io.IOException: java.util.concurrent.ExecutionException: org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/.regioninfo could only be written to 2 of the 3 required nodes for RS-3-2-1024k. There are 3 datanode(s) running and 3 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2480) at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:293) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3075) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:932) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:603) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.hbase.util.ModifyRegionUtils.createRegions(ModifyRegionUtils.java:141) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.ModifyRegionUtils.createRegions(ModifyRegionUtils.java:101) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.master.procedure.CreateTableProcedure$1.createHdfsRegions(CreateTableProcedure.java:360) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.createFsLayout(CreateTableProcedure.java:379) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.createFsLayout(CreateTableProcedure.java:353) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.executeFromState(CreateTableProcedure.java:107) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.executeFromState(CreateTableProcedure.java:59) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:188) ~[hbase-procedure-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:941) ~[hbase-procedure-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1825) ~[hbase-procedure-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.executeProcedure(ProcedureExecutor.java:1503) ~[hbase-procedure-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.runProcedure(ProcedureExecutor.java:2154) ~[hbase-procedure-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2181) ~[hbase-procedure-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] Caused by: java.util.concurrent.ExecutionException: org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/.regioninfo could only be written to 2 of the 3 required nodes for RS-3-2-1024k. There are 3 datanode(s) running and 3 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2480) at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:293) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3075) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:932) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:603) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at java.util.concurrent.FutureTask.report(FutureTask.java:122) ~[?:?] at java.util.concurrent.FutureTask.get(FutureTask.java:191) ~[?:?] at org.apache.hadoop.hbase.util.ModifyRegionUtils.createRegions(ModifyRegionUtils.java:135) ~[classes/:4.0.0-alpha-1-SNAPSHOT] ... 13 more Caused by: org.apache.hadoop.ipc.RemoteException: File /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/.regioninfo could only be written to 2 of the 3 required nodes for RS-3-2-1024k. There are 3 datanode(s) running and 3 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2480) at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:293) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3075) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:932) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:603) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$addBlock$11(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DFSOutputStream.addBlock(DFSOutputStream.java:1143) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.allocateNewBlock(DFSStripedOutputStream.java:508) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.writeChunk(DFSStripedOutputStream.java:561) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks(FSOutputSummer.java:220) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:165) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:146) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.closeImpl(DFSStripedOutputStream.java:1234) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:861) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:77) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:106) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoFileContent(HRegionFileSystem.java:815) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoOnFilesystem(HRegionFileSystem.java:906) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoOnFilesystem(HRegionFileSystem.java:868) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.createRegionOnFileSystem(HRegionFileSystem.java:936) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.createRegionDir(HRegion.java:7606) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:7574) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:7557) ~[classes/:?] at org.apache.hadoop.hbase.util.ModifyRegionUtils.createRegion(ModifyRegionUtils.java:162) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.ModifyRegionUtils$1.call(ModifyRegionUtils.java:128) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.ModifyRegionUtils$1.call(ModifyRegionUtils.java:125) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) ~[?:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:20,716 INFO [PEWorker-3 {}] procedure2.TimeoutExecutorThread(82): ADDED pid=4, state=WAITING_TIMEOUT:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC; timeout=1015, timestamp=1732477521731 2024-11-24T19:45:20,793 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:21,003 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:21,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775740_1008 (size=1189) 2024-11-24T19:45:21,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_-9223372036854775773_1004 (size=42) 2024-11-24T19:45:21,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_-9223372036854775741_1008 (size=1189) 2024-11-24T19:45:21,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775772_1004 (size=42) 2024-11-24T19:45:21,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775693_1015 (size=32) 2024-11-24T19:45:21,248 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_-9223372036854775692_1015 (size=32) 2024-11-24T19:45:21,249 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775757_1006 (size=196) 2024-11-24T19:45:21,313 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:21,737 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-24T19:45:21,745 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:21,748 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(153): Directory hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3 empty. 2024-11-24T19:45:21,749 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:21,749 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(313): Archived TestHBaseWalOnEC regions 2024-11-24T19:45:21,758 WARN [PEWorker-4 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:21,758 WARN [PEWorker-4 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:21,761 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:36866 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36866 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:21,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775664_1023 (size=392) 2024-11-24T19:45:21,766 WARN [PEWorker-4 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:21,768 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 005eb91681f1bb20289d5dd15edd33d3, NAME => 'TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9 2024-11-24T19:45:21,773 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:21,773 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:21,777 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:36880 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36880 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:21,782 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775648_1025 (size=51) 2024-11-24T19:45:21,783 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:21,783 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:21,783 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 005eb91681f1bb20289d5dd15edd33d3, disabling compactions & flushes 2024-11-24T19:45:21,783 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:21,783 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:21,783 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. after waiting 0 ms 2024-11-24T19:45:21,783 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:21,783 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:21,784 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 005eb91681f1bb20289d5dd15edd33d3: Waiting for close lock at 1732477521783Disabling compacts and flushes for region at 1732477521783Disabling writes for close at 1732477521783Writing region close event to WAL at 1732477521783Closed at 1732477521783 2024-11-24T19:45:21,786 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-11-24T19:45:21,792 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1732477521786"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1732477521786"}]},"ts":"1732477521786"} 2024-11-24T19:45:21,797 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-11-24T19:45:21,798 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-11-24T19:45:21,801 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1732477521799"}]},"ts":"1732477521799"} 2024-11-24T19:45:21,805 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-11-24T19:45:21,806 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(204): Hosts are {32a03c886245=0} racks are {/default-rack=0} 2024-11-24T19:45:21,807 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-24T19:45:21,807 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-24T19:45:21,808 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-24T19:45:21,808 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-24T19:45:21,808 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-24T19:45:21,808 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-24T19:45:21,808 INFO [PEWorker-4 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-24T19:45:21,808 INFO [PEWorker-4 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-24T19:45:21,808 INFO [PEWorker-4 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-24T19:45:21,808 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-24T19:45:21,809 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=005eb91681f1bb20289d5dd15edd33d3, ASSIGN}] 2024-11-24T19:45:21,813 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=005eb91681f1bb20289d5dd15edd33d3, ASSIGN 2024-11-24T19:45:21,815 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=005eb91681f1bb20289d5dd15edd33d3, ASSIGN; state=OFFLINE, location=32a03c886245,37807,1732477517489; forceNewPlan=false, retain=false 2024-11-24T19:45:21,822 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:21,968 INFO [32a03c886245:33203 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-11-24T19:45:21,969 INFO [PEWorker-2 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=005eb91681f1bb20289d5dd15edd33d3, regionState=OPENING, regionLocation=32a03c886245,37807,1732477517489 2024-11-24T19:45:21,974 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=005eb91681f1bb20289d5dd15edd33d3, ASSIGN because future has completed 2024-11-24T19:45:21,986 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 005eb91681f1bb20289d5dd15edd33d3, server=32a03c886245,37807,1732477517489}] 2024-11-24T19:45:22,149 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-24T19:45:22,151 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35189, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-24T19:45:22,157 INFO [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:22,157 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 005eb91681f1bb20289d5dd15edd33d3, NAME => 'TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3.', STARTKEY => '', ENDKEY => ''} 2024-11-24T19:45:22,157 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,157 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:22,158 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,158 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,160 INFO [StoreOpener-005eb91681f1bb20289d5dd15edd33d3-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,163 INFO [StoreOpener-005eb91681f1bb20289d5dd15edd33d3-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 005eb91681f1bb20289d5dd15edd33d3 columnFamilyName cf 2024-11-24T19:45:22,163 DEBUG [StoreOpener-005eb91681f1bb20289d5dd15edd33d3-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:22,164 INFO [StoreOpener-005eb91681f1bb20289d5dd15edd33d3-1 {}] regionserver.HStore(327): Store=005eb91681f1bb20289d5dd15edd33d3/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-24T19:45:22,164 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,166 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,167 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,168 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,168 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,170 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,179 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-24T19:45:22,180 INFO [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 005eb91681f1bb20289d5dd15edd33d3; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67268979, jitterRate=0.0023858994245529175}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-11-24T19:45:22,180 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:22,181 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 005eb91681f1bb20289d5dd15edd33d3: Running coprocessor pre-open hook at 1732477522158Writing region info on filesystem at 1732477522158Initializing all the Stores at 1732477522160 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477522160Cleaning up temporary data from old regions at 1732477522168 (+8 ms)Running coprocessor post-open hooks at 1732477522180 (+12 ms)Region opened successfully at 1732477522181 (+1 ms) 2024-11-24T19:45:22,183 INFO [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3., pid=6, masterSystemTime=1732477522148 2024-11-24T19:45:22,186 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:22,187 INFO [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:22,188 INFO [PEWorker-4 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=005eb91681f1bb20289d5dd15edd33d3, regionState=OPEN, openSeqNum=2, regionLocation=32a03c886245,37807,1732477517489 2024-11-24T19:45:22,192 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 005eb91681f1bb20289d5dd15edd33d3, server=32a03c886245,37807,1732477517489 because future has completed 2024-11-24T19:45:22,199 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-11-24T19:45:22,200 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 005eb91681f1bb20289d5dd15edd33d3, server=32a03c886245,37807,1732477517489 in 208 msec 2024-11-24T19:45:22,203 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-11-24T19:45:22,204 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=005eb91681f1bb20289d5dd15edd33d3, ASSIGN in 390 msec 2024-11-24T19:45:22,205 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-11-24T19:45:22,205 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1732477522205"}]},"ts":"1732477522205"} 2024-11-24T19:45:22,208 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-11-24T19:45:22,210 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-11-24T19:45:22,213 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 1.5490 sec 2024-11-24T19:45:22,833 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:22,834 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-11-24T19:45:22,834 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-11-24T19:45:22,835 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-24T19:45:22,842 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-11-24T19:45:22,843 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-24T19:45:22,844 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-11-24T19:45:22,857 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3., hostname=32a03c886245,37807,1732477517489, seqNum=2] 2024-11-24T19:45:22,858 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-24T19:45:22,860 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46710, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-24T19:45:22,867 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-11-24T19:45:22,872 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-11-24T19:45:22,874 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-24T19:45:22,874 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-11-24T19:45:22,876 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-11-24T19:45:22,878 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-11-24T19:45:22,983 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-24T19:45:23,045 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37807 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-11-24T19:45:23,046 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:23,051 INFO [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 005eb91681f1bb20289d5dd15edd33d3 1/1 column families, dataSize=32 B heapSize=360 B 2024-11-24T19:45:23,101 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/.tmp/cf/c7a9825c94b04e0da95023e5b8ddfdaa is 36, key is row/cf:cq/1732477522861/Put/seqid=0 2024-11-24T19:45:23,107 WARN [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,107 WARN [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,111 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-344286536_22 at /127.0.0.1:39420 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:37841:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39420 dst: /127.0.0.1:37841 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:23,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_-9223372036854775632_1027 (size=4787) 2024-11-24T19:45:23,116 WARN [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:23,117 INFO [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/.tmp/cf/c7a9825c94b04e0da95023e5b8ddfdaa 2024-11-24T19:45:23,154 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/.tmp/cf/c7a9825c94b04e0da95023e5b8ddfdaa as hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/cf/c7a9825c94b04e0da95023e5b8ddfdaa 2024-11-24T19:45:23,164 INFO [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/cf/c7a9825c94b04e0da95023e5b8ddfdaa, entries=1, sequenceid=5, filesize=4.7 K 2024-11-24T19:45:23,171 INFO [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 005eb91681f1bb20289d5dd15edd33d3 in 120ms, sequenceid=5, compaction requested=false 2024-11-24T19:45:23,172 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-11-24T19:45:23,175 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 005eb91681f1bb20289d5dd15edd33d3: 2024-11-24T19:45:23,175 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:23,177 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-11-24T19:45:23,179 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-11-24T19:45:23,185 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-11-24T19:45:23,185 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 303 msec 2024-11-24T19:45:23,188 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 317 msec 2024-11-24T19:45:23,192 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=33203 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-24T19:45:23,192 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-11-24T19:45:23,204 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-11-24T19:45:23,204 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-24T19:45:23,204 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:23,208 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,209 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,209 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-11-24T19:45:23,209 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-11-24T19:45:23,209 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=372944084, stopped=false 2024-11-24T19:45:23,209 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=32a03c886245,33203,1732477516430 2024-11-24T19:45:23,262 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:23,262 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:23,262 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:23,262 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:23,262 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-24T19:45:23,262 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:23,262 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:23,262 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:23,262 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:23,263 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-24T19:45:23,264 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:23,264 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,264 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:23,264 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:23,264 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:23,264 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:23,265 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32a03c886245,39441,1732477517291' ***** 2024-11-24T19:45:23,266 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-24T19:45:23,266 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32a03c886245,38025,1732477517418' ***** 2024-11-24T19:45:23,266 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-24T19:45:23,266 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32a03c886245,37807,1732477517489' ***** 2024-11-24T19:45:23,266 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-24T19:45:23,266 INFO [RS:0;32a03c886245:39441 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-24T19:45:23,267 INFO [RS:2;32a03c886245:37807 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-24T19:45:23,267 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-24T19:45:23,267 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-24T19:45:23,268 INFO [RS:1;32a03c886245:38025 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-24T19:45:23,268 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-24T19:45:23,268 INFO [RS:1;32a03c886245:38025 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-24T19:45:23,268 INFO [RS:2;32a03c886245:37807 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-24T19:45:23,268 INFO [RS:0;32a03c886245:39441 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-24T19:45:23,268 INFO [RS:0;32a03c886245:39441 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-24T19:45:23,268 INFO [RS:1;32a03c886245:38025 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-24T19:45:23,268 INFO [RS:2;32a03c886245:37807 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-24T19:45:23,269 INFO [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(959): stopping server 32a03c886245,39441,1732477517291 2024-11-24T19:45:23,269 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(959): stopping server 32a03c886245,38025,1732477517418 2024-11-24T19:45:23,269 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(3091): Received CLOSE for 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:23,269 INFO [RS:0;32a03c886245:39441 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-24T19:45:23,269 INFO [RS:1;32a03c886245:38025 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-24T19:45:23,269 INFO [RS:0;32a03c886245:39441 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;32a03c886245:39441. 2024-11-24T19:45:23,269 INFO [RS:1;32a03c886245:38025 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;32a03c886245:38025. 2024-11-24T19:45:23,269 DEBUG [RS:0;32a03c886245:39441 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:23,269 DEBUG [RS:0;32a03c886245:39441 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,269 DEBUG [RS:1;32a03c886245:38025 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:23,269 DEBUG [RS:1;32a03c886245:38025 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,269 INFO [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(976): stopping server 32a03c886245,39441,1732477517291; all regions closed. 2024-11-24T19:45:23,269 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(959): stopping server 32a03c886245,37807,1732477517489 2024-11-24T19:45:23,269 INFO [RS:1;32a03c886245:38025 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-24T19:45:23,269 INFO [RS:2;32a03c886245:37807 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-24T19:45:23,269 INFO [RS:1;32a03c886245:38025 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-24T19:45:23,269 INFO [RS:1;32a03c886245:38025 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-24T19:45:23,269 INFO [RS:2;32a03c886245:37807 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;32a03c886245:37807. 2024-11-24T19:45:23,270 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-11-24T19:45:23,270 DEBUG [RS:2;32a03c886245:37807 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:23,270 DEBUG [RS:2;32a03c886245:37807 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,270 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 005eb91681f1bb20289d5dd15edd33d3, disabling compactions & flushes 2024-11-24T19:45:23,270 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-24T19:45:23,270 INFO [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:23,270 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:23,270 DEBUG [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1325): Online Regions={005eb91681f1bb20289d5dd15edd33d3=TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3.} 2024-11-24T19:45:23,270 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-24T19:45:23,270 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. after waiting 0 ms 2024-11-24T19:45:23,270 DEBUG [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-11-24T19:45:23,270 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:23,270 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-24T19:45:23,270 DEBUG [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1351): Waiting on 005eb91681f1bb20289d5dd15edd33d3 2024-11-24T19:45:23,270 DEBUG [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-11-24T19:45:23,270 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-24T19:45:23,270 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-24T19:45:23,271 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-24T19:45:23,271 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-24T19:45:23,271 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-11-24T19:45:23,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_1073741827_1017 (size=93) 2024-11-24T19:45:23,275 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_1073741827_1017 (size=93) 2024-11-24T19:45:23,283 DEBUG [RS:0;32a03c886245:39441 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/oldWALs 2024-11-24T19:45:23,283 INFO [RS:0;32a03c886245:39441 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 32a03c886245%2C39441%2C1732477517291:(num 1732477519695) 2024-11-24T19:45:23,283 DEBUG [RS:0;32a03c886245:39441 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,283 INFO [RS:0;32a03c886245:39441 {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:23,283 INFO [RS:0;32a03c886245:39441 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-24T19:45:23,283 INFO [RS:0;32a03c886245:39441 {}] hbase.ChoreService(370): Chore service for: regionserver/32a03c886245:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-11-24T19:45:23,283 INFO [RS:0;32a03c886245:39441 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-24T19:45:23,284 INFO [RS:0;32a03c886245:39441 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-24T19:45:23,284 INFO [regionserver/32a03c886245:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-24T19:45:23,284 INFO [RS:0;32a03c886245:39441 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-24T19:45:23,284 INFO [RS:0;32a03c886245:39441 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-24T19:45:23,284 INFO [RS:0;32a03c886245:39441 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39441 2024-11-24T19:45:23,288 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/default/TestHBaseWalOnEC/005eb91681f1bb20289d5dd15edd33d3/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-11-24T19:45:23,290 INFO [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:23,290 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 005eb91681f1bb20289d5dd15edd33d3: Waiting for close lock at 1732477523269Running coprocessor pre-close hooks at 1732477523270 (+1 ms)Disabling compacts and flushes for region at 1732477523270Disabling writes for close at 1732477523270Writing region close event to WAL at 1732477523271 (+1 ms)Running coprocessor post-close hooks at 1732477523289 (+18 ms)Closed at 1732477523290 (+1 ms) 2024-11-24T19:45:23,291 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3. 2024-11-24T19:45:23,297 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32a03c886245,39441,1732477517291 2024-11-24T19:45:23,297 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-24T19:45:23,297 INFO [RS:0;32a03c886245:39441 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-24T19:45:23,306 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32a03c886245,39441,1732477517291] 2024-11-24T19:45:23,306 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/info/bcd2fae716564ebdad8585285dea4f91 is 153, key is TestHBaseWalOnEC,,1732477520652.005eb91681f1bb20289d5dd15edd33d3./info:regioninfo/1732477522188/Put/seqid=0 2024-11-24T19:45:23,309 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,309 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,314 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32a03c886245,39441,1732477517291 already deleted, retry=false 2024-11-24T19:45:23,314 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1689614638_22 at /127.0.0.1:37222 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37222 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:23,314 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32a03c886245,39441,1732477517291 expired; onlineServers=2 2024-11-24T19:45:23,318 INFO [regionserver/32a03c886245:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:23,320 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775616_1029 (size=6637) 2024-11-24T19:45:23,321 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:23,321 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/info/bcd2fae716564ebdad8585285dea4f91 2024-11-24T19:45:23,325 INFO [regionserver/32a03c886245:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:23,330 INFO [regionserver/32a03c886245:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:23,349 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/ns/4e9fdfebd99d4bb2b3385abc4ab53367 is 43, key is default/ns:d/1732477520392/Put/seqid=0 2024-11-24T19:45:23,352 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,352 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,355 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1689614638_22 at /127.0.0.1:37252 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37252 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:23,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775600_1031 (size=5153) 2024-11-24T19:45:23,360 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:23,360 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/ns/4e9fdfebd99d4bb2b3385abc4ab53367 2024-11-24T19:45:23,388 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/table/cae9d5640d1347a789d9f7af6aa7d391 is 52, key is TestHBaseWalOnEC/table:state/1732477522205/Put/seqid=0 2024-11-24T19:45:23,390 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,390 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,393 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1689614638_22 at /127.0.0.1:37274 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:43851:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37274 dst: /127.0.0.1:43851 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:23,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_-9223372036854775584_1033 (size=5249) 2024-11-24T19:45:23,398 WARN [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:23,398 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/table/cae9d5640d1347a789d9f7af6aa7d391 2024-11-24T19:45:23,406 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:23,406 DEBUG [pool-65-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39441-0x1016faa8a030001, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:23,406 INFO [RS:0;32a03c886245:39441 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-24T19:45:23,407 INFO [RS:0;32a03c886245:39441 {}] regionserver.HRegionServer(1031): Exiting; stopping=32a03c886245,39441,1732477517291; zookeeper connection closed. 2024-11-24T19:45:23,407 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4e0064fe {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4e0064fe 2024-11-24T19:45:23,409 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/info/bcd2fae716564ebdad8585285dea4f91 as hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/info/bcd2fae716564ebdad8585285dea4f91 2024-11-24T19:45:23,418 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/info/bcd2fae716564ebdad8585285dea4f91, entries=10, sequenceid=11, filesize=6.5 K 2024-11-24T19:45:23,420 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/ns/4e9fdfebd99d4bb2b3385abc4ab53367 as hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/ns/4e9fdfebd99d4bb2b3385abc4ab53367 2024-11-24T19:45:23,428 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/ns/4e9fdfebd99d4bb2b3385abc4ab53367, entries=2, sequenceid=11, filesize=5.0 K 2024-11-24T19:45:23,430 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/.tmp/table/cae9d5640d1347a789d9f7af6aa7d391 as hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/table/cae9d5640d1347a789d9f7af6aa7d391 2024-11-24T19:45:23,438 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/table/cae9d5640d1347a789d9f7af6aa7d391, entries=2, sequenceid=11, filesize=5.1 K 2024-11-24T19:45:23,440 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 169ms, sequenceid=11, compaction requested=false 2024-11-24T19:45:23,440 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-11-24T19:45:23,448 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-11-24T19:45:23,449 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-24T19:45:23,450 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-24T19:45:23,450 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1732477523270Running coprocessor pre-close hooks at 1732477523270Disabling compacts and flushes for region at 1732477523270Disabling writes for close at 1732477523271 (+1 ms)Obtaining lock to block concurrent updates at 1732477523271Preparing flush snapshotting stores in 1588230740 at 1732477523271Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1732477523272 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1732477523273 (+1 ms)Flushing 1588230740/info: creating writer at 1732477523273Flushing 1588230740/info: appending metadata at 1732477523301 (+28 ms)Flushing 1588230740/info: closing flushed file at 1732477523301Flushing 1588230740/ns: creating writer at 1732477523332 (+31 ms)Flushing 1588230740/ns: appending metadata at 1732477523348 (+16 ms)Flushing 1588230740/ns: closing flushed file at 1732477523348Flushing 1588230740/table: creating writer at 1732477523368 (+20 ms)Flushing 1588230740/table: appending metadata at 1732477523385 (+17 ms)Flushing 1588230740/table: closing flushed file at 1732477523385Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@62f3a5d2: reopening flushed file at 1732477523407 (+22 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6c6b576: reopening flushed file at 1732477523418 (+11 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6fcdda0b: reopening flushed file at 1732477523429 (+11 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 169ms, sequenceid=11, compaction requested=false at 1732477523440 (+11 ms)Writing region close event to WAL at 1732477523442 (+2 ms)Running coprocessor post-close hooks at 1732477523449 (+7 ms)Closed at 1732477523449 2024-11-24T19:45:23,450 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-11-24T19:45:23,470 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(976): stopping server 32a03c886245,37807,1732477517489; all regions closed. 2024-11-24T19:45:23,470 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(976): stopping server 32a03c886245,38025,1732477517418; all regions closed. 2024-11-24T19:45:23,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_1073741826_1016 (size=1298) 2024-11-24T19:45:23,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_1073741826_1016 (size=1298) 2024-11-24T19:45:23,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_1073741829_1019 (size=2751) 2024-11-24T19:45:23,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43851 is added to blk_1073741826_1016 (size=1298) 2024-11-24T19:45:23,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_1073741829_1019 (size=2751) 2024-11-24T19:45:23,480 DEBUG [RS:1;32a03c886245:38025 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/oldWALs 2024-11-24T19:45:23,480 INFO [RS:1;32a03c886245:38025 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 32a03c886245%2C38025%2C1732477517418.meta:.meta(num 1732477520117) 2024-11-24T19:45:23,480 DEBUG [RS:2;32a03c886245:37807 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/oldWALs 2024-11-24T19:45:23,480 INFO [RS:2;32a03c886245:37807 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 32a03c886245%2C37807%2C1732477517489:(num 1732477519694) 2024-11-24T19:45:23,480 DEBUG [RS:2;32a03c886245:37807 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,480 INFO [RS:2;32a03c886245:37807 {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:23,480 INFO [RS:2;32a03c886245:37807 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-24T19:45:23,480 INFO [RS:2;32a03c886245:37807 {}] hbase.ChoreService(370): Chore service for: regionserver/32a03c886245:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-24T19:45:23,481 INFO [RS:2;32a03c886245:37807 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-24T19:45:23,481 INFO [RS:2;32a03c886245:37807 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-24T19:45:23,481 INFO [regionserver/32a03c886245:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-24T19:45:23,481 INFO [RS:2;32a03c886245:37807 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-24T19:45:23,481 INFO [RS:2;32a03c886245:37807 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-24T19:45:23,481 INFO [RS:2;32a03c886245:37807 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:37807 2024-11-24T19:45:23,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_1073741828_1018 (size=93) 2024-11-24T19:45:23,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_1073741828_1018 (size=93) 2024-11-24T19:45:23,487 DEBUG [RS:1;32a03c886245:38025 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/oldWALs 2024-11-24T19:45:23,487 INFO [RS:1;32a03c886245:38025 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 32a03c886245%2C38025%2C1732477517418:(num 1732477519694) 2024-11-24T19:45:23,487 DEBUG [RS:1;32a03c886245:38025 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:23,487 INFO [RS:1;32a03c886245:38025 {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:23,487 INFO [RS:1;32a03c886245:38025 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-24T19:45:23,487 INFO [RS:1;32a03c886245:38025 {}] hbase.ChoreService(370): Chore service for: regionserver/32a03c886245:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-24T19:45:23,487 INFO [RS:1;32a03c886245:38025 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-24T19:45:23,488 INFO [regionserver/32a03c886245:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-24T19:45:23,488 INFO [RS:1;32a03c886245:38025 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:38025 2024-11-24T19:45:23,511 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32a03c886245,37807,1732477517489 2024-11-24T19:45:23,511 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-24T19:45:23,511 INFO [RS:2;32a03c886245:37807 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-24T19:45:23,519 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32a03c886245,38025,1732477517418 2024-11-24T19:45:23,519 INFO [RS:1;32a03c886245:38025 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-24T19:45:23,528 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32a03c886245,37807,1732477517489] 2024-11-24T19:45:23,544 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32a03c886245,37807,1732477517489 already deleted, retry=false 2024-11-24T19:45:23,544 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32a03c886245,37807,1732477517489 expired; onlineServers=1 2024-11-24T19:45:23,544 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32a03c886245,38025,1732477517418] 2024-11-24T19:45:23,555 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32a03c886245,38025,1732477517418 already deleted, retry=false 2024-11-24T19:45:23,556 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32a03c886245,38025,1732477517418 expired; onlineServers=0 2024-11-24T19:45:23,556 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '32a03c886245,33203,1732477516430' ***** 2024-11-24T19:45:23,556 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-11-24T19:45:23,556 INFO [M:0;32a03c886245:33203 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-24T19:45:23,557 INFO [M:0;32a03c886245:33203 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-24T19:45:23,557 DEBUG [M:0;32a03c886245:33203 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-11-24T19:45:23,557 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-11-24T19:45:23,557 DEBUG [M:0;32a03c886245:33203 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-11-24T19:45:23,557 DEBUG [master/32a03c886245:0:becomeActiveMaster-HFileCleaner.small.0-1732477519223 {}] cleaner.HFileCleaner(306): Exit Thread[master/32a03c886245:0:becomeActiveMaster-HFileCleaner.small.0-1732477519223,5,FailOnTimeoutGroup] 2024-11-24T19:45:23,557 DEBUG [master/32a03c886245:0:becomeActiveMaster-HFileCleaner.large.0-1732477519212 {}] cleaner.HFileCleaner(306): Exit Thread[master/32a03c886245:0:becomeActiveMaster-HFileCleaner.large.0-1732477519212,5,FailOnTimeoutGroup] 2024-11-24T19:45:23,558 INFO [M:0;32a03c886245:33203 {}] hbase.ChoreService(370): Chore service for: master/32a03c886245:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-11-24T19:45:23,558 INFO [M:0;32a03c886245:33203 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-24T19:45:23,559 DEBUG [M:0;32a03c886245:33203 {}] master.HMaster(1795): Stopping service threads 2024-11-24T19:45:23,559 INFO [M:0;32a03c886245:33203 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-11-24T19:45:23,559 INFO [M:0;32a03c886245:33203 {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-24T19:45:23,560 ERROR [M:0;32a03c886245:33203 {}] procedure2.ProcedureExecutor(763): There are still active thread in group java.lang.ThreadGroup[name=PEWorkerGroup,maxpri=10], see STDOUT java.lang.ThreadGroup[name=PEWorkerGroup,maxpri=10] Thread[HFileArchiver-1,5,PEWorkerGroup] 2024-11-24T19:45:23,561 INFO [M:0;32a03c886245:33203 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-11-24T19:45:23,561 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-11-24T19:45:23,569 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-11-24T19:45:23,569 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:23,569 DEBUG [M:0;32a03c886245:33203 {}] zookeeper.ZKUtil(347): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-11-24T19:45:23,570 WARN [M:0;32a03c886245:33203 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-11-24T19:45:23,571 INFO [M:0;32a03c886245:33203 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/.lastflushedseqids 2024-11-24T19:45:23,578 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,578 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,581 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:48256 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:35937:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48256 dst: /127.0.0.1:35937 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:23,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775568_1035 (size=127) 2024-11-24T19:45:23,586 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:23,586 INFO [M:0;32a03c886245:33203 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-11-24T19:45:23,587 INFO [M:0;32a03c886245:33203 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-11-24T19:45:23,587 DEBUG [M:0;32a03c886245:33203 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-24T19:45:23,587 INFO [M:0;32a03c886245:33203 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:23,587 DEBUG [M:0;32a03c886245:33203 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:23,587 DEBUG [M:0;32a03c886245:33203 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-24T19:45:23,587 DEBUG [M:0;32a03c886245:33203 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:23,587 INFO [M:0;32a03c886245:33203 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=28.27 KB heapSize=35.71 KB 2024-11-24T19:45:23,605 DEBUG [M:0;32a03c886245:33203 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a2ce69a279f841c68601156f145d436d is 82, key is hbase:meta,,1/info:regioninfo/1732477520218/Put/seqid=0 2024-11-24T19:45:23,607 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,607 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,610 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:48268 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:35937:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48268 dst: /127.0.0.1:35937 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:23,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775552_1037 (size=5672) 2024-11-24T19:45:23,614 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:23,614 INFO [M:0;32a03c886245:33203 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=74 (bloomFilter=true), to=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a2ce69a279f841c68601156f145d436d 2024-11-24T19:45:23,628 INFO [RS:2;32a03c886245:37807 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-24T19:45:23,628 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:23,628 INFO [RS:2;32a03c886245:37807 {}] regionserver.HRegionServer(1031): Exiting; stopping=32a03c886245,37807,1732477517489; zookeeper connection closed. 2024-11-24T19:45:23,628 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37807-0x1016faa8a030003, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:23,629 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@49f11400 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@49f11400 2024-11-24T19:45:23,636 INFO [RS:1;32a03c886245:38025 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-24T19:45:23,636 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:23,636 INFO [RS:1;32a03c886245:38025 {}] regionserver.HRegionServer(1031): Exiting; stopping=32a03c886245,38025,1732477517418; zookeeper connection closed. 2024-11-24T19:45:23,636 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38025-0x1016faa8a030002, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:23,637 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7a0931a7 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7a0931a7 2024-11-24T19:45:23,637 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-11-24T19:45:23,638 DEBUG [M:0;32a03c886245:33203 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ef32ec90b42540ddb4aaae2d252d5e28 is 751, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1732477522212/Put/seqid=0 2024-11-24T19:45:23,640 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,640 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,643 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:48286 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:35937:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48286 dst: /127.0.0.1:35937 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:23,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775536_1039 (size=6442) 2024-11-24T19:45:23,647 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:23,648 INFO [M:0;32a03c886245:33203 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=27.59 KB at sequenceid=74 (bloomFilter=true), to=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ef32ec90b42540ddb4aaae2d252d5e28 2024-11-24T19:45:23,670 DEBUG [M:0;32a03c886245:33203 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/37759a429e414df08c60c623e0117f98 is 69, key is 32a03c886245,37807,1732477517489/rs:state/1732477519292/Put/seqid=0 2024-11-24T19:45:23,672 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,672 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-24T19:45:23,675 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1180939187_22 at /127.0.0.1:48300 [Receiving block BP-29170651-172.17.0.2-1732477511872:blk_-9223372036854775520_1040] {}] datanode.DataXceiver(331): 127.0.0.1:35937:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48300 dst: /127.0.0.1:35937 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-24T19:45:23,678 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_-9223372036854775520_1041 (size=5294) 2024-11-24T19:45:23,679 WARN [M:0;32a03c886245:33203 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-24T19:45:23,679 INFO [M:0;32a03c886245:33203 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=74 (bloomFilter=true), to=hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/37759a429e414df08c60c623e0117f98 2024-11-24T19:45:23,687 DEBUG [M:0;32a03c886245:33203 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a2ce69a279f841c68601156f145d436d as hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/a2ce69a279f841c68601156f145d436d 2024-11-24T19:45:23,695 INFO [M:0;32a03c886245:33203 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/a2ce69a279f841c68601156f145d436d, entries=8, sequenceid=74, filesize=5.5 K 2024-11-24T19:45:23,697 DEBUG [M:0;32a03c886245:33203 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ef32ec90b42540ddb4aaae2d252d5e28 as hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ef32ec90b42540ddb4aaae2d252d5e28 2024-11-24T19:45:23,706 INFO [M:0;32a03c886245:33203 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ef32ec90b42540ddb4aaae2d252d5e28, entries=8, sequenceid=74, filesize=6.3 K 2024-11-24T19:45:23,708 DEBUG [M:0;32a03c886245:33203 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/37759a429e414df08c60c623e0117f98 as hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/37759a429e414df08c60c623e0117f98 2024-11-24T19:45:23,716 INFO [M:0;32a03c886245:33203 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/37759a429e414df08c60c623e0117f98, entries=3, sequenceid=74, filesize=5.2 K 2024-11-24T19:45:23,718 INFO [M:0;32a03c886245:33203 {}] regionserver.HRegion(3140): Finished flush of dataSize ~28.27 KB/28947, heapSize ~35.41 KB/36264, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 130ms, sequenceid=74, compaction requested=false 2024-11-24T19:45:23,719 INFO [M:0;32a03c886245:33203 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:23,719 DEBUG [M:0;32a03c886245:33203 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1732477523587Disabling compacts and flushes for region at 1732477523587Disabling writes for close at 1732477523587Obtaining lock to block concurrent updates at 1732477523587Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1732477523587Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=28947, getHeapSize=36504, getOffHeapSize=0, getCellsCount=87 at 1732477523588 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1732477523589 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1732477523589Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1732477523605 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1732477523605Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1732477523621 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1732477523638 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1732477523638Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1732477523655 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1732477523670 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1732477523670Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6797e905: reopening flushed file at 1732477523686 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5f4c613e: reopening flushed file at 1732477523696 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@328d32d: reopening flushed file at 1732477523706 (+10 ms)Finished flush of dataSize ~28.27 KB/28947, heapSize ~35.41 KB/36264, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 130ms, sequenceid=74, compaction requested=false at 1732477523718 (+12 ms)Writing region close event to WAL at 1732477523719 (+1 ms)Closed at 1732477523719 2024-11-24T19:45:23,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35937 is added to blk_1073741825_1011 (size=34278) 2024-11-24T19:45:23,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37841 is added to blk_1073741825_1011 (size=34278) 2024-11-24T19:45:23,724 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-24T19:45:23,724 INFO [M:0;32a03c886245:33203 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-11-24T19:45:23,724 INFO [M:0;32a03c886245:33203 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:33203 2024-11-24T19:45:23,724 INFO [M:0;32a03c886245:33203 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-24T19:45:23,856 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:23,856 INFO [M:0;32a03c886245:33203 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-24T19:45:23,856 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:33203-0x1016faa8a030000, quorum=127.0.0.1:61658, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:23,909 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@17f8e572{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:23,914 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11aaae40{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-24T19:45:23,915 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-24T19:45:23,915 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@14c1b227{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-24T19:45:23,915 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@28d0ee11{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir/,STOPPED} 2024-11-24T19:45:23,917 WARN [BP-29170651-172.17.0.2-1732477511872 heartbeating to localhost/127.0.0.1:40555 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-24T19:45:23,917 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-24T19:45:23,917 WARN [BP-29170651-172.17.0.2-1732477511872 heartbeating to localhost/127.0.0.1:40555 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-29170651-172.17.0.2-1732477511872 (Datanode Uuid 2a332ccd-6d46-413a-8a7e-6880197ec257) service to localhost/127.0.0.1:40555 2024-11-24T19:45:23,917 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-24T19:45:23,918 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data5/current/BP-29170651-172.17.0.2-1732477511872 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:23,918 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data6/current/BP-29170651-172.17.0.2-1732477511872 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:23,919 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-24T19:45:23,921 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@133f1bad{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:23,921 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@39860596{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-24T19:45:23,922 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-24T19:45:23,922 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@125705fb{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-24T19:45:23,922 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@50510811{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir/,STOPPED} 2024-11-24T19:45:23,923 WARN [BP-29170651-172.17.0.2-1732477511872 heartbeating to localhost/127.0.0.1:40555 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-24T19:45:23,923 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-24T19:45:23,923 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-24T19:45:23,923 WARN [BP-29170651-172.17.0.2-1732477511872 heartbeating to localhost/127.0.0.1:40555 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-29170651-172.17.0.2-1732477511872 (Datanode Uuid 0277df7e-84cc-4e77-988c-6586260f6c5b) service to localhost/127.0.0.1:40555 2024-11-24T19:45:23,924 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data3/current/BP-29170651-172.17.0.2-1732477511872 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:23,924 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data4/current/BP-29170651-172.17.0.2-1732477511872 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:23,924 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-24T19:45:23,930 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3990ff75{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:23,930 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@48731e1b{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-24T19:45:23,930 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-24T19:45:23,931 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4f02cc61{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-24T19:45:23,931 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@179ed6d0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir/,STOPPED} 2024-11-24T19:45:23,933 WARN [BP-29170651-172.17.0.2-1732477511872 heartbeating to localhost/127.0.0.1:40555 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-24T19:45:23,933 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-24T19:45:23,933 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-24T19:45:23,933 WARN [BP-29170651-172.17.0.2-1732477511872 heartbeating to localhost/127.0.0.1:40555 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-29170651-172.17.0.2-1732477511872 (Datanode Uuid 665164a6-ecc0-44fd-8c67-f432dcd8b1f3) service to localhost/127.0.0.1:40555 2024-11-24T19:45:23,933 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data1/current/BP-29170651-172.17.0.2-1732477511872 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:23,934 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/cluster_696c92ed-5464-2487-1e6d-2fb45b032620/data/data2/current/BP-29170651-172.17.0.2-1732477511872 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:23,934 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-24T19:45:23,943 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5cb83937{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-24T19:45:23,944 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@69b5b273{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-24T19:45:23,944 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-24T19:45:23,944 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3cbd6fd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-24T19:45:23,944 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@b03c34d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir/,STOPPED} 2024-11-24T19:45:23,952 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-11-24T19:45:23,981 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-11-24T19:45:23,987 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=87 (was 156), OpenFileDescriptor=433 (was 393) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=330 (was 291) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=10813 (was 11172) 2024-11-24T19:45:23,993 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=87, OpenFileDescriptor=433, MaxFileDescriptor=1048576, SystemLoadAverage=330, ProcessCount=11, AvailableMemoryMB=10813 2024-11-24T19:45:23,993 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.log.dir so I do NOT create it in target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/f9236714-226d-9519-0759-d0883e3572d6/hadoop.tmp.dir so I do NOT create it in target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c, deleteOnExit=true 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/test.cache.data in system properties and HBase conf 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.tmp.dir in system properties and HBase conf 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir in system properties and HBase conf 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/mapreduce.cluster.local.dir in system properties and HBase conf 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-11-24T19:45:23,994 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-11-24T19:45:23,995 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-11-24T19:45:23,995 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-11-24T19:45:23,995 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-11-24T19:45:23,995 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-11-24T19:45:23,995 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-24T19:45:23,995 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/nfs.dump.dir in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/java.io.tmpdir in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-11-24T19:45:23,996 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-11-24T19:45:24,288 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:24,294 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-24T19:45:24,298 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-24T19:45:24,298 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-24T19:45:24,299 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-24T19:45:24,299 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:24,300 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4c686d6e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir/,AVAILABLE} 2024-11-24T19:45:24,300 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5574eaf2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-24T19:45:24,424 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5609bdf8{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/java.io.tmpdir/jetty-localhost-33643-hadoop-hdfs-3_4_1-tests_jar-_-any-14161001832137682760/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-24T19:45:24,425 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@7cbf28d2{HTTP/1.1, (http/1.1)}{localhost:33643} 2024-11-24T19:45:24,425 INFO [Time-limited test {}] server.Server(415): Started @14884ms 2024-11-24T19:45:24,686 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:24,689 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-24T19:45:24,690 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-24T19:45:24,690 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-24T19:45:24,690 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-24T19:45:24,690 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6c411bc1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir/,AVAILABLE} 2024-11-24T19:45:24,691 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@d63bfae{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-24T19:45:24,785 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@484ab650{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/java.io.tmpdir/jetty-localhost-43819-hadoop-hdfs-3_4_1-tests_jar-_-any-3953826704591983050/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:24,785 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1d09d7d5{HTTP/1.1, (http/1.1)}{localhost:43819} 2024-11-24T19:45:24,785 INFO [Time-limited test {}] server.Server(415): Started @15243ms 2024-11-24T19:45:24,786 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-24T19:45:24,817 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:24,820 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-24T19:45:24,820 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-24T19:45:24,820 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-24T19:45:24,820 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-24T19:45:24,821 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@17cd1c9f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir/,AVAILABLE} 2024-11-24T19:45:24,822 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2b2d9a02{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-24T19:45:24,915 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@569dbdc8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/java.io.tmpdir/jetty-localhost-33531-hadoop-hdfs-3_4_1-tests_jar-_-any-3234583350966275805/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:24,915 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@18c4cb78{HTTP/1.1, (http/1.1)}{localhost:33531} 2024-11-24T19:45:24,916 INFO [Time-limited test {}] server.Server(415): Started @15374ms 2024-11-24T19:45:24,917 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-24T19:45:24,942 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-24T19:45:24,945 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-24T19:45:24,946 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-24T19:45:24,946 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-24T19:45:24,946 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-24T19:45:24,946 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@25de4fb7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir/,AVAILABLE} 2024-11-24T19:45:24,947 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@498db938{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-24T19:45:25,112 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@51eb0253{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/java.io.tmpdir/jetty-localhost-44857-hadoop-hdfs-3_4_1-tests_jar-_-any-16518033758414492467/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:25,112 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@38066021{HTTP/1.1, (http/1.1)}{localhost:44857} 2024-11-24T19:45:25,112 INFO [Time-limited test {}] server.Server(415): Started @15570ms 2024-11-24T19:45:25,113 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-24T19:45:25,163 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-24T19:45:25,163 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-24T19:45:25,668 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-11-24T19:45:25,679 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-24T19:45:25,727 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-24T19:45:25,727 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-24T19:45:25,937 WARN [Thread-566 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data1/current/BP-41452304-172.17.0.2-1732477524022/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:25,937 WARN [Thread-567 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data2/current/BP-41452304-172.17.0.2-1732477524022/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:25,959 WARN [Thread-507 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-24T19:45:25,962 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xfbec51d351fe8a72 with lease ID 0x90871196d7529fba: Processing first storage report for DS-d1ad34a2-2a04-44b7-a31e-884d53c201e4 from datanode DatanodeRegistration(127.0.0.1:36477, datanodeUuid=854a62d0-6f39-4749-a4be-2b3ae7e2c987, infoPort=37763, infoSecurePort=0, ipcPort=42049, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022) 2024-11-24T19:45:25,962 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xfbec51d351fe8a72 with lease ID 0x90871196d7529fba: from storage DS-d1ad34a2-2a04-44b7-a31e-884d53c201e4 node DatanodeRegistration(127.0.0.1:36477, datanodeUuid=854a62d0-6f39-4749-a4be-2b3ae7e2c987, infoPort=37763, infoSecurePort=0, ipcPort=42049, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:25,962 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xfbec51d351fe8a72 with lease ID 0x90871196d7529fba: Processing first storage report for DS-e917297a-edca-4dbd-a558-07ec0b14c244 from datanode DatanodeRegistration(127.0.0.1:36477, datanodeUuid=854a62d0-6f39-4749-a4be-2b3ae7e2c987, infoPort=37763, infoSecurePort=0, ipcPort=42049, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022) 2024-11-24T19:45:25,962 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xfbec51d351fe8a72 with lease ID 0x90871196d7529fba: from storage DS-e917297a-edca-4dbd-a558-07ec0b14c244 node DatanodeRegistration(127.0.0.1:36477, datanodeUuid=854a62d0-6f39-4749-a4be-2b3ae7e2c987, infoPort=37763, infoSecurePort=0, ipcPort=42049, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:26,247 WARN [Thread-579 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data3/current/BP-41452304-172.17.0.2-1732477524022/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:26,248 WARN [Thread-580 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data4/current/BP-41452304-172.17.0.2-1732477524022/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:26,263 WARN [Thread-530 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-24T19:45:26,265 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x493d087e3a98d273 with lease ID 0x90871196d7529fbb: Processing first storage report for DS-b68736b5-faf2-4bf4-8d1e-507b8f19e592 from datanode DatanodeRegistration(127.0.0.1:34183, datanodeUuid=fd9037f1-a98d-4491-9463-e5df8ca942c3, infoPort=44217, infoSecurePort=0, ipcPort=40003, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022) 2024-11-24T19:45:26,265 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x493d087e3a98d273 with lease ID 0x90871196d7529fbb: from storage DS-b68736b5-faf2-4bf4-8d1e-507b8f19e592 node DatanodeRegistration(127.0.0.1:34183, datanodeUuid=fd9037f1-a98d-4491-9463-e5df8ca942c3, infoPort=44217, infoSecurePort=0, ipcPort=40003, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:26,265 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x493d087e3a98d273 with lease ID 0x90871196d7529fbb: Processing first storage report for DS-43f31f07-eda0-4f7e-a67e-4396591eda30 from datanode DatanodeRegistration(127.0.0.1:34183, datanodeUuid=fd9037f1-a98d-4491-9463-e5df8ca942c3, infoPort=44217, infoSecurePort=0, ipcPort=40003, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022) 2024-11-24T19:45:26,265 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x493d087e3a98d273 with lease ID 0x90871196d7529fbb: from storage DS-43f31f07-eda0-4f7e-a67e-4396591eda30 node DatanodeRegistration(127.0.0.1:34183, datanodeUuid=fd9037f1-a98d-4491-9463-e5df8ca942c3, infoPort=44217, infoSecurePort=0, ipcPort=40003, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:26,271 WARN [Thread-591 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data6/current/BP-41452304-172.17.0.2-1732477524022/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:26,271 WARN [Thread-590 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data5/current/BP-41452304-172.17.0.2-1732477524022/current, will proceed with Du for space computation calculation, 2024-11-24T19:45:26,288 WARN [Thread-552 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-24T19:45:26,291 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2d564fcdb015a822 with lease ID 0x90871196d7529fbc: Processing first storage report for DS-f28f74a3-0c7c-4bd4-a241-e1f7d43dce8a from datanode DatanodeRegistration(127.0.0.1:36373, datanodeUuid=de0a7a9d-20f5-40b1-a5e6-85b077b5d204, infoPort=45169, infoSecurePort=0, ipcPort=42113, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022) 2024-11-24T19:45:26,291 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2d564fcdb015a822 with lease ID 0x90871196d7529fbc: from storage DS-f28f74a3-0c7c-4bd4-a241-e1f7d43dce8a node DatanodeRegistration(127.0.0.1:36373, datanodeUuid=de0a7a9d-20f5-40b1-a5e6-85b077b5d204, infoPort=45169, infoSecurePort=0, ipcPort=42113, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-24T19:45:26,291 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2d564fcdb015a822 with lease ID 0x90871196d7529fbc: Processing first storage report for DS-b48fa623-1e96-4a76-8848-1b8f223f5adc from datanode DatanodeRegistration(127.0.0.1:36373, datanodeUuid=de0a7a9d-20f5-40b1-a5e6-85b077b5d204, infoPort=45169, infoSecurePort=0, ipcPort=42113, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022) 2024-11-24T19:45:26,291 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2d564fcdb015a822 with lease ID 0x90871196d7529fbc: from storage DS-b48fa623-1e96-4a76-8848-1b8f223f5adc node DatanodeRegistration(127.0.0.1:36373, datanodeUuid=de0a7a9d-20f5-40b1-a5e6-85b077b5d204, infoPort=45169, infoSecurePort=0, ipcPort=42113, storageInfo=lv=-57;cid=testClusterID;nsid=649704057;c=1732477524022), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-24T19:45:26,362 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3 2024-11-24T19:45:26,367 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/zookeeper_0, clientPort=55315, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-11-24T19:45:26,368 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=55315 2024-11-24T19:45:26,368 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,369 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,380 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741825_1001 (size=7) 2024-11-24T19:45:26,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741825_1001 (size=7) 2024-11-24T19:45:26,382 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741825_1001 (size=7) 2024-11-24T19:45:26,383 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448 with version=8 2024-11-24T19:45:26,383 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:40555/user/jenkins/test-data/82fff9a3-c6b4-0ee7-40b3-62d7b00459c9/hbase-staging 2024-11-24T19:45:26,385 INFO [Time-limited test {}] client.ConnectionUtils(128): master/32a03c886245:0 server-side Connection retries=45 2024-11-24T19:45:26,385 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,385 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,385 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-24T19:45:26,385 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,385 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-24T19:45:26,386 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-11-24T19:45:26,386 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-24T19:45:26,386 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36267 2024-11-24T19:45:26,388 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:36267 connecting to ZooKeeper ensemble=127.0.0.1:55315 2024-11-24T19:45:26,438 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:362670x0, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-24T19:45:26,439 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:36267-0x1016faab4230000 connected 2024-11-24T19:45:26,540 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,544 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,548 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:26,549 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448, hbase.cluster.distributed=false 2024-11-24T19:45:26,553 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-24T19:45:26,553 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36267 2024-11-24T19:45:26,554 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36267 2024-11-24T19:45:26,554 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36267 2024-11-24T19:45:26,554 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36267 2024-11-24T19:45:26,555 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36267 2024-11-24T19:45:26,570 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32a03c886245:0 server-side Connection retries=45 2024-11-24T19:45:26,570 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,570 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,570 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-24T19:45:26,570 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,570 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-24T19:45:26,570 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-24T19:45:26,571 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-24T19:45:26,571 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:44439 2024-11-24T19:45:26,572 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:44439 connecting to ZooKeeper ensemble=127.0.0.1:55315 2024-11-24T19:45:26,573 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,575 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,585 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:444390x0, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-24T19:45:26,586 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:44439-0x1016faab4230001 connected 2024-11-24T19:45:26,586 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:26,586 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-24T19:45:26,587 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-24T19:45:26,587 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-24T19:45:26,588 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-24T19:45:26,589 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=44439 2024-11-24T19:45:26,589 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=44439 2024-11-24T19:45:26,590 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=44439 2024-11-24T19:45:26,590 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=44439 2024-11-24T19:45:26,590 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=44439 2024-11-24T19:45:26,607 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32a03c886245:0 server-side Connection retries=45 2024-11-24T19:45:26,607 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,607 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,607 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-24T19:45:26,607 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,607 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-24T19:45:26,607 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-24T19:45:26,607 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-24T19:45:26,608 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:35081 2024-11-24T19:45:26,609 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:35081 connecting to ZooKeeper ensemble=127.0.0.1:55315 2024-11-24T19:45:26,610 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,612 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,622 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:350810x0, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-24T19:45:26,622 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:350810x0, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:26,622 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35081-0x1016faab4230002 connected 2024-11-24T19:45:26,623 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-24T19:45:26,623 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-24T19:45:26,624 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-24T19:45:26,625 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-24T19:45:26,626 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35081 2024-11-24T19:45:26,626 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35081 2024-11-24T19:45:26,627 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35081 2024-11-24T19:45:26,627 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35081 2024-11-24T19:45:26,627 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35081 2024-11-24T19:45:26,644 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/32a03c886245:0 server-side Connection retries=45 2024-11-24T19:45:26,644 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,644 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,644 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-24T19:45:26,644 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-24T19:45:26,644 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-24T19:45:26,644 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-24T19:45:26,645 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-24T19:45:26,645 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:45177 2024-11-24T19:45:26,647 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:45177 connecting to ZooKeeper ensemble=127.0.0.1:55315 2024-11-24T19:45:26,647 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,648 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,660 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:451770x0, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-24T19:45:26,661 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:451770x0, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:26,661 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45177-0x1016faab4230003 connected 2024-11-24T19:45:26,661 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-24T19:45:26,662 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-24T19:45:26,663 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-24T19:45:26,664 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-24T19:45:26,665 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45177 2024-11-24T19:45:26,665 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45177 2024-11-24T19:45:26,665 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45177 2024-11-24T19:45:26,666 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45177 2024-11-24T19:45:26,666 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45177 2024-11-24T19:45:26,678 DEBUG [M:0;32a03c886245:36267 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;32a03c886245:36267 2024-11-24T19:45:26,678 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/32a03c886245,36267,1732477526385 2024-11-24T19:45:26,688 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:26,688 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:26,688 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:26,688 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:26,689 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/32a03c886245,36267,1732477526385 2024-11-24T19:45:26,697 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-24T19:45:26,697 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-24T19:45:26,697 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-24T19:45:26,697 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,697 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,697 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,697 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,697 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-11-24T19:45:26,698 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/32a03c886245,36267,1732477526385 from backup master directory 2024-11-24T19:45:26,705 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:26,705 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:26,705 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:26,705 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/32a03c886245,36267,1732477526385 2024-11-24T19:45:26,705 WARN [master/32a03c886245:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-24T19:45:26,705 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-24T19:45:26,705 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=32a03c886245,36267,1732477526385 2024-11-24T19:45:26,712 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/hbase.id] with ID: 416c9059-ca21-4b55-9b45-3d71930a7e06 2024-11-24T19:45:26,712 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/.tmp/hbase.id 2024-11-24T19:45:26,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741826_1002 (size=42) 2024-11-24T19:45:26,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741826_1002 (size=42) 2024-11-24T19:45:26,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741826_1002 (size=42) 2024-11-24T19:45:26,722 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/.tmp/hbase.id]:[hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/hbase.id] 2024-11-24T19:45:26,740 INFO [master/32a03c886245:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-24T19:45:26,740 INFO [master/32a03c886245:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-11-24T19:45:26,742 INFO [master/32a03c886245:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-11-24T19:45:26,752 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,752 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,752 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,752 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,761 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741827_1003 (size=196) 2024-11-24T19:45:26,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741827_1003 (size=196) 2024-11-24T19:45:26,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741827_1003 (size=196) 2024-11-24T19:45:26,763 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-24T19:45:26,764 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-11-24T19:45:26,765 INFO [master/32a03c886245:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-24T19:45:26,777 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741828_1004 (size=1189) 2024-11-24T19:45:26,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741828_1004 (size=1189) 2024-11-24T19:45:26,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741828_1004 (size=1189) 2024-11-24T19:45:26,779 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store 2024-11-24T19:45:26,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741829_1005 (size=34) 2024-11-24T19:45:26,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741829_1005 (size=34) 2024-11-24T19:45:26,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741829_1005 (size=34) 2024-11-24T19:45:26,789 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:26,789 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-24T19:45:26,789 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:26,789 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:26,789 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-24T19:45:26,789 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:26,789 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:26,789 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1732477526789Disabling compacts and flushes for region at 1732477526789Disabling writes for close at 1732477526789Writing region close event to WAL at 1732477526789Closed at 1732477526789 2024-11-24T19:45:26,791 WARN [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/.initializing 2024-11-24T19:45:26,791 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/WALs/32a03c886245,36267,1732477526385 2024-11-24T19:45:26,795 INFO [master/32a03c886245:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C36267%2C1732477526385, suffix=, logDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/WALs/32a03c886245,36267,1732477526385, archiveDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/oldWALs, maxLogs=10 2024-11-24T19:45:26,796 INFO [master/32a03c886245:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 32a03c886245%2C36267%2C1732477526385.1732477526796 2024-11-24T19:45:26,807 INFO [master/32a03c886245:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/WALs/32a03c886245,36267,1732477526385/32a03c886245%2C36267%2C1732477526385.1732477526796 2024-11-24T19:45:26,810 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45169:45169),(127.0.0.1/127.0.0.1:44217:44217),(127.0.0.1/127.0.0.1:37763:37763)] 2024-11-24T19:45:26,811 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-11-24T19:45:26,812 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:26,812 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,812 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,814 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,816 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-11-24T19:45:26,816 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:26,817 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:26,817 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,819 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-11-24T19:45:26,819 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:26,819 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-24T19:45:26,819 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,822 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-11-24T19:45:26,822 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:26,822 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-24T19:45:26,823 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,824 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-11-24T19:45:26,824 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:26,825 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-24T19:45:26,825 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,826 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,826 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,828 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,828 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,829 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-24T19:45:26,830 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-11-24T19:45:26,833 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-24T19:45:26,834 INFO [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61312749, jitterRate=-0.08636884391307831}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-24T19:45:26,835 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1732477526812Initializing all the Stores at 1732477526813 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477526813Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477526814 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477526814Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477526814Cleaning up temporary data from old regions at 1732477526828 (+14 ms)Region opened successfully at 1732477526835 (+7 ms) 2024-11-24T19:45:26,835 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-11-24T19:45:26,840 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4ad37c6c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32a03c886245/172.17.0.2:0 2024-11-24T19:45:26,841 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-11-24T19:45:26,841 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-11-24T19:45:26,841 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-11-24T19:45:26,841 INFO [master/32a03c886245:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-11-24T19:45:26,842 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-11-24T19:45:26,842 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-11-24T19:45:26,842 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-11-24T19:45:26,845 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-11-24T19:45:26,846 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-11-24T19:45:26,859 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-11-24T19:45:26,860 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-11-24T19:45:26,872 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-11-24T19:45:26,872 INFO [master/32a03c886245:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-11-24T19:45:26,873 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-11-24T19:45:26,880 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-11-24T19:45:26,881 INFO [master/32a03c886245:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-11-24T19:45:26,882 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-11-24T19:45:26,893 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-11-24T19:45:26,895 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-11-24T19:45:26,905 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-11-24T19:45:26,908 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-11-24T19:45:26,918 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-11-24T19:45:26,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:26,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:26,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:26,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:26,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,930 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,931 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=32a03c886245,36267,1732477526385, sessionid=0x1016faab4230000, setting cluster-up flag (Was=false) 2024-11-24T19:45:26,947 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,947 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,947 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,947 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,972 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-11-24T19:45:26,973 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=32a03c886245,36267,1732477526385 2024-11-24T19:45:26,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:26,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:27,021 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-11-24T19:45:27,023 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=32a03c886245,36267,1732477526385 2024-11-24T19:45:27,024 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-11-24T19:45:27,027 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-11-24T19:45:27,027 INFO [master/32a03c886245:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-11-24T19:45:27,027 INFO [master/32a03c886245:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-11-24T19:45:27,028 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 32a03c886245,36267,1732477526385 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-11-24T19:45:27,030 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/32a03c886245:0, corePoolSize=5, maxPoolSize=5 2024-11-24T19:45:27,030 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/32a03c886245:0, corePoolSize=5, maxPoolSize=5 2024-11-24T19:45:27,030 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/32a03c886245:0, corePoolSize=5, maxPoolSize=5 2024-11-24T19:45:27,030 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/32a03c886245:0, corePoolSize=5, maxPoolSize=5 2024-11-24T19:45:27,030 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/32a03c886245:0, corePoolSize=10, maxPoolSize=10 2024-11-24T19:45:27,030 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,030 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/32a03c886245:0, corePoolSize=2, maxPoolSize=2 2024-11-24T19:45:27,031 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,032 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1732477557032 2024-11-24T19:45:27,032 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-11-24T19:45:27,032 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-11-24T19:45:27,032 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-11-24T19:45:27,032 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-11-24T19:45:27,032 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-11-24T19:45:27,032 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-11-24T19:45:27,033 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,033 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-11-24T19:45:27,033 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-11-24T19:45:27,033 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-11-24T19:45:27,034 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-24T19:45:27,034 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-11-24T19:45:27,034 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-11-24T19:45:27,034 INFO [master/32a03c886245:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-11-24T19:45:27,035 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/32a03c886245:0:becomeActiveMaster-HFileCleaner.large.0-1732477527035,5,FailOnTimeoutGroup] 2024-11-24T19:45:27,035 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/32a03c886245:0:becomeActiveMaster-HFileCleaner.small.0-1732477527035,5,FailOnTimeoutGroup] 2024-11-24T19:45:27,035 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,035 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-11-24T19:45:27,036 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,036 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,036 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,036 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-11-24T19:45:27,049 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741831_1007 (size=1321) 2024-11-24T19:45:27,049 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741831_1007 (size=1321) 2024-11-24T19:45:27,050 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741831_1007 (size=1321) 2024-11-24T19:45:27,051 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-11-24T19:45:27,051 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448 2024-11-24T19:45:27,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741832_1008 (size=32) 2024-11-24T19:45:27,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741832_1008 (size=32) 2024-11-24T19:45:27,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741832_1008 (size=32) 2024-11-24T19:45:27,063 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:27,065 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-24T19:45:27,067 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-24T19:45:27,067 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,068 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:27,068 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-24T19:45:27,068 INFO [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(746): ClusterId : 416c9059-ca21-4b55-9b45-3d71930a7e06 2024-11-24T19:45:27,068 INFO [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(746): ClusterId : 416c9059-ca21-4b55-9b45-3d71930a7e06 2024-11-24T19:45:27,068 DEBUG [RS:1;32a03c886245:35081 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-24T19:45:27,068 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(746): ClusterId : 416c9059-ca21-4b55-9b45-3d71930a7e06 2024-11-24T19:45:27,068 DEBUG [RS:2;32a03c886245:45177 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-24T19:45:27,068 DEBUG [RS:0;32a03c886245:44439 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-24T19:45:27,070 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-24T19:45:27,070 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,070 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:27,070 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-24T19:45:27,072 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-24T19:45:27,072 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,073 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:27,073 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-24T19:45:27,074 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-24T19:45:27,075 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,075 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:27,075 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-24T19:45:27,076 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740 2024-11-24T19:45:27,077 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740 2024-11-24T19:45:27,078 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-24T19:45:27,078 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-24T19:45:27,079 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-24T19:45:27,080 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-24T19:45:27,081 DEBUG [RS:1;32a03c886245:35081 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-24T19:45:27,081 DEBUG [RS:2;32a03c886245:45177 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-24T19:45:27,081 DEBUG [RS:0;32a03c886245:44439 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-24T19:45:27,081 DEBUG [RS:2;32a03c886245:45177 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-24T19:45:27,081 DEBUG [RS:1;32a03c886245:35081 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-24T19:45:27,081 DEBUG [RS:0;32a03c886245:44439 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-24T19:45:27,083 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-24T19:45:27,084 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70900186, jitterRate=0.056495100259780884}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-24T19:45:27,084 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1732477527063Initializing all the Stores at 1732477527064 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477527064Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477527065 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477527065Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477527065Cleaning up temporary data from old regions at 1732477527078 (+13 ms)Region opened successfully at 1732477527084 (+6 ms) 2024-11-24T19:45:27,084 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-24T19:45:27,085 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-24T19:45:27,085 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-24T19:45:27,085 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-24T19:45:27,085 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-24T19:45:27,085 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-24T19:45:27,085 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1732477527084Disabling compacts and flushes for region at 1732477527084Disabling writes for close at 1732477527085 (+1 ms)Writing region close event to WAL at 1732477527085Closed at 1732477527085 2024-11-24T19:45:27,087 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-24T19:45:27,087 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-11-24T19:45:27,087 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-11-24T19:45:27,089 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-24T19:45:27,089 DEBUG [RS:2;32a03c886245:45177 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-24T19:45:27,089 DEBUG [RS:1;32a03c886245:35081 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-24T19:45:27,089 DEBUG [RS:1;32a03c886245:35081 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@26468eee, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32a03c886245/172.17.0.2:0 2024-11-24T19:45:27,089 DEBUG [RS:2;32a03c886245:45177 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2f055c6e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32a03c886245/172.17.0.2:0 2024-11-24T19:45:27,090 DEBUG [RS:0;32a03c886245:44439 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-24T19:45:27,090 DEBUG [RS:0;32a03c886245:44439 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2548db2e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=32a03c886245/172.17.0.2:0 2024-11-24T19:45:27,090 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-11-24T19:45:27,100 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;32a03c886245:44439 2024-11-24T19:45:27,100 DEBUG [RS:1;32a03c886245:35081 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;32a03c886245:35081 2024-11-24T19:45:27,100 DEBUG [RS:2;32a03c886245:45177 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;32a03c886245:45177 2024-11-24T19:45:27,100 INFO [RS:0;32a03c886245:44439 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-24T19:45:27,100 INFO [RS:0;32a03c886245:44439 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-24T19:45:27,100 INFO [RS:2;32a03c886245:45177 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-24T19:45:27,100 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-24T19:45:27,100 INFO [RS:1;32a03c886245:35081 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-24T19:45:27,100 INFO [RS:2;32a03c886245:45177 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-24T19:45:27,100 INFO [RS:1;32a03c886245:35081 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-24T19:45:27,100 DEBUG [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-24T19:45:27,100 DEBUG [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-24T19:45:27,101 INFO [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(2659): reportForDuty to master=32a03c886245,36267,1732477526385 with port=45177, startcode=1732477526644 2024-11-24T19:45:27,101 INFO [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(2659): reportForDuty to master=32a03c886245,36267,1732477526385 with port=35081, startcode=1732477526607 2024-11-24T19:45:27,101 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(2659): reportForDuty to master=32a03c886245,36267,1732477526385 with port=44439, startcode=1732477526570 2024-11-24T19:45:27,101 DEBUG [RS:2;32a03c886245:45177 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-24T19:45:27,101 DEBUG [RS:1;32a03c886245:35081 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-24T19:45:27,101 DEBUG [RS:0;32a03c886245:44439 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-24T19:45:27,103 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58141, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-11-24T19:45:27,104 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34169, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-11-24T19:45:27,104 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56929, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-11-24T19:45:27,104 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36267 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32a03c886245,44439,1732477526570 2024-11-24T19:45:27,104 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36267 {}] master.ServerManager(517): Registering regionserver=32a03c886245,44439,1732477526570 2024-11-24T19:45:27,106 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36267 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32a03c886245,35081,1732477526607 2024-11-24T19:45:27,106 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36267 {}] master.ServerManager(517): Registering regionserver=32a03c886245,35081,1732477526607 2024-11-24T19:45:27,106 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448 2024-11-24T19:45:27,107 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42003 2024-11-24T19:45:27,107 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-24T19:45:27,108 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36267 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 32a03c886245,45177,1732477526644 2024-11-24T19:45:27,108 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36267 {}] master.ServerManager(517): Registering regionserver=32a03c886245,45177,1732477526644 2024-11-24T19:45:27,108 DEBUG [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448 2024-11-24T19:45:27,109 DEBUG [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42003 2024-11-24T19:45:27,109 DEBUG [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-24T19:45:27,110 DEBUG [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448 2024-11-24T19:45:27,111 DEBUG [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42003 2024-11-24T19:45:27,111 DEBUG [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-24T19:45:27,118 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-24T19:45:27,151 DEBUG [RS:0;32a03c886245:44439 {}] zookeeper.ZKUtil(111): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32a03c886245,44439,1732477526570 2024-11-24T19:45:27,152 WARN [RS:0;32a03c886245:44439 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-24T19:45:27,152 INFO [RS:0;32a03c886245:44439 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-24T19:45:27,152 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32a03c886245,44439,1732477526570] 2024-11-24T19:45:27,152 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32a03c886245,45177,1732477526644] 2024-11-24T19:45:27,152 DEBUG [RS:1;32a03c886245:35081 {}] zookeeper.ZKUtil(111): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32a03c886245,35081,1732477526607 2024-11-24T19:45:27,152 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [32a03c886245,35081,1732477526607] 2024-11-24T19:45:27,152 WARN [RS:1;32a03c886245:35081 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-24T19:45:27,152 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,44439,1732477526570 2024-11-24T19:45:27,152 DEBUG [RS:2;32a03c886245:45177 {}] zookeeper.ZKUtil(111): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/32a03c886245,45177,1732477526644 2024-11-24T19:45:27,152 INFO [RS:1;32a03c886245:35081 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-24T19:45:27,152 WARN [RS:2;32a03c886245:45177 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-24T19:45:27,152 DEBUG [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,35081,1732477526607 2024-11-24T19:45:27,153 INFO [RS:2;32a03c886245:45177 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-24T19:45:27,153 DEBUG [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,45177,1732477526644 2024-11-24T19:45:27,160 INFO [RS:1;32a03c886245:35081 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-24T19:45:27,160 INFO [RS:2;32a03c886245:45177 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-24T19:45:27,160 INFO [RS:0;32a03c886245:44439 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-24T19:45:27,163 INFO [RS:2;32a03c886245:45177 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-24T19:45:27,164 INFO [RS:2;32a03c886245:45177 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-24T19:45:27,164 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,164 INFO [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-24T19:45:27,165 INFO [RS:2;32a03c886245:45177 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-24T19:45:27,165 INFO [RS:1;32a03c886245:35081 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-24T19:45:27,165 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,165 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 INFO [RS:1;32a03c886245:35081 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-24T19:45:27,166 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 INFO [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32a03c886245:0, corePoolSize=2, maxPoolSize=2 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,166 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,167 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:27,167 DEBUG [RS:2;32a03c886245:45177 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:27,167 INFO [RS:1;32a03c886245:35081 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-24T19:45:27,167 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,167 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,167 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,167 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,167 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,167 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,167 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32a03c886245:0, corePoolSize=2, maxPoolSize=2 2024-11-24T19:45:27,167 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,167 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,168 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,168 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,168 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,168 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,168 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:27,168 DEBUG [RS:1;32a03c886245:35081 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:27,170 INFO [RS:0;32a03c886245:44439 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-24T19:45:27,171 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:0;32a03c886245:44439 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-24T19:45:27,171 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,45177,1732477526644-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,171 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,35081,1732477526607-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-24T19:45:27,173 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-24T19:45:27,174 INFO [RS:0;32a03c886245:44439 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-24T19:45:27,174 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,174 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/32a03c886245:0, corePoolSize=2, maxPoolSize=2 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,175 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/32a03c886245:0, corePoolSize=1, maxPoolSize=1 2024-11-24T19:45:27,176 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:27,176 DEBUG [RS:0;32a03c886245:44439 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0, corePoolSize=3, maxPoolSize=3 2024-11-24T19:45:27,177 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,177 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,177 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,177 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,177 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,177 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,44439,1732477526570-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-24T19:45:27,186 INFO [RS:2;32a03c886245:45177 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-24T19:45:27,186 INFO [RS:1;32a03c886245:35081 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-24T19:45:27,186 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,45177,1732477526644-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,186 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,35081,1732477526607-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,186 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,186 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,186 INFO [RS:2;32a03c886245:45177 {}] regionserver.Replication(171): 32a03c886245,45177,1732477526644 started 2024-11-24T19:45:27,186 INFO [RS:1;32a03c886245:35081 {}] regionserver.Replication(171): 32a03c886245,35081,1732477526607 started 2024-11-24T19:45:27,190 INFO [RS:0;32a03c886245:44439 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-24T19:45:27,190 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,44439,1732477526570-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,190 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,190 INFO [RS:0;32a03c886245:44439 {}] regionserver.Replication(171): 32a03c886245,44439,1732477526570 started 2024-11-24T19:45:27,199 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,199 INFO [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(1482): Serving as 32a03c886245,45177,1732477526644, RpcServer on 32a03c886245/172.17.0.2:45177, sessionid=0x1016faab4230003 2024-11-24T19:45:27,199 DEBUG [RS:2;32a03c886245:45177 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-24T19:45:27,199 DEBUG [RS:2;32a03c886245:45177 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32a03c886245,45177,1732477526644 2024-11-24T19:45:27,199 DEBUG [RS:2;32a03c886245:45177 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,45177,1732477526644' 2024-11-24T19:45:27,199 DEBUG [RS:2;32a03c886245:45177 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-24T19:45:27,200 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,200 INFO [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(1482): Serving as 32a03c886245,35081,1732477526607, RpcServer on 32a03c886245/172.17.0.2:35081, sessionid=0x1016faab4230002 2024-11-24T19:45:27,200 DEBUG [RS:1;32a03c886245:35081 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-24T19:45:27,200 DEBUG [RS:1;32a03c886245:35081 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32a03c886245,35081,1732477526607 2024-11-24T19:45:27,200 DEBUG [RS:1;32a03c886245:35081 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,35081,1732477526607' 2024-11-24T19:45:27,200 DEBUG [RS:1;32a03c886245:35081 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-24T19:45:27,200 DEBUG [RS:2;32a03c886245:45177 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-24T19:45:27,200 DEBUG [RS:1;32a03c886245:35081 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-24T19:45:27,201 DEBUG [RS:2;32a03c886245:45177 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-24T19:45:27,201 DEBUG [RS:2;32a03c886245:45177 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-24T19:45:27,201 DEBUG [RS:2;32a03c886245:45177 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32a03c886245,45177,1732477526644 2024-11-24T19:45:27,201 DEBUG [RS:2;32a03c886245:45177 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,45177,1732477526644' 2024-11-24T19:45:27,201 DEBUG [RS:2;32a03c886245:45177 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-24T19:45:27,201 DEBUG [RS:1;32a03c886245:35081 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-24T19:45:27,201 DEBUG [RS:1;32a03c886245:35081 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-24T19:45:27,201 DEBUG [RS:1;32a03c886245:35081 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32a03c886245,35081,1732477526607 2024-11-24T19:45:27,201 DEBUG [RS:1;32a03c886245:35081 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,35081,1732477526607' 2024-11-24T19:45:27,201 DEBUG [RS:1;32a03c886245:35081 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-24T19:45:27,201 DEBUG [RS:2;32a03c886245:45177 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-24T19:45:27,201 DEBUG [RS:1;32a03c886245:35081 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-24T19:45:27,201 DEBUG [RS:2;32a03c886245:45177 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-24T19:45:27,201 INFO [RS:2;32a03c886245:45177 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-24T19:45:27,202 INFO [RS:2;32a03c886245:45177 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-24T19:45:27,202 DEBUG [RS:1;32a03c886245:35081 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-24T19:45:27,202 INFO [RS:1;32a03c886245:35081 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-24T19:45:27,202 INFO [RS:1;32a03c886245:35081 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-24T19:45:27,203 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,203 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1482): Serving as 32a03c886245,44439,1732477526570, RpcServer on 32a03c886245/172.17.0.2:44439, sessionid=0x1016faab4230001 2024-11-24T19:45:27,203 DEBUG [RS:0;32a03c886245:44439 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-24T19:45:27,203 DEBUG [RS:0;32a03c886245:44439 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 32a03c886245,44439,1732477526570 2024-11-24T19:45:27,203 DEBUG [RS:0;32a03c886245:44439 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,44439,1732477526570' 2024-11-24T19:45:27,203 DEBUG [RS:0;32a03c886245:44439 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-24T19:45:27,204 DEBUG [RS:0;32a03c886245:44439 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-24T19:45:27,204 DEBUG [RS:0;32a03c886245:44439 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-24T19:45:27,204 DEBUG [RS:0;32a03c886245:44439 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-24T19:45:27,204 DEBUG [RS:0;32a03c886245:44439 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 32a03c886245,44439,1732477526570 2024-11-24T19:45:27,204 DEBUG [RS:0;32a03c886245:44439 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '32a03c886245,44439,1732477526570' 2024-11-24T19:45:27,204 DEBUG [RS:0;32a03c886245:44439 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-24T19:45:27,205 DEBUG [RS:0;32a03c886245:44439 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-24T19:45:27,205 DEBUG [RS:0;32a03c886245:44439 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-24T19:45:27,205 INFO [RS:0;32a03c886245:44439 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-24T19:45:27,205 INFO [RS:0;32a03c886245:44439 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-24T19:45:27,241 WARN [32a03c886245:36267 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-11-24T19:45:27,308 INFO [RS:1;32a03c886245:35081 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C35081%2C1732477526607, suffix=, logDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,35081,1732477526607, archiveDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/oldWALs, maxLogs=32 2024-11-24T19:45:27,309 INFO [RS:2;32a03c886245:45177 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C45177%2C1732477526644, suffix=, logDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,45177,1732477526644, archiveDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/oldWALs, maxLogs=32 2024-11-24T19:45:27,311 INFO [RS:0;32a03c886245:44439 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C44439%2C1732477526570, suffix=, logDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,44439,1732477526570, archiveDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/oldWALs, maxLogs=32 2024-11-24T19:45:27,311 INFO [RS:1;32a03c886245:35081 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 32a03c886245%2C35081%2C1732477526607.1732477527311 2024-11-24T19:45:27,312 INFO [RS:2;32a03c886245:45177 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 32a03c886245%2C45177%2C1732477526644.1732477527312 2024-11-24T19:45:27,313 INFO [RS:0;32a03c886245:44439 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 32a03c886245%2C44439%2C1732477526570.1732477527313 2024-11-24T19:45:27,327 INFO [RS:1;32a03c886245:35081 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,35081,1732477526607/32a03c886245%2C35081%2C1732477526607.1732477527311 2024-11-24T19:45:27,331 INFO [RS:2;32a03c886245:45177 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,45177,1732477526644/32a03c886245%2C45177%2C1732477526644.1732477527312 2024-11-24T19:45:27,332 DEBUG [RS:1;32a03c886245:35081 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:44217:44217),(127.0.0.1/127.0.0.1:45169:45169),(127.0.0.1/127.0.0.1:37763:37763)] 2024-11-24T19:45:27,332 INFO [RS:0;32a03c886245:44439 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,44439,1732477526570/32a03c886245%2C44439%2C1732477526570.1732477527313 2024-11-24T19:45:27,333 DEBUG [RS:2;32a03c886245:45177 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45169:45169),(127.0.0.1/127.0.0.1:44217:44217),(127.0.0.1/127.0.0.1:37763:37763)] 2024-11-24T19:45:27,339 DEBUG [RS:0;32a03c886245:44439 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45169:45169),(127.0.0.1/127.0.0.1:37763:37763),(127.0.0.1/127.0.0.1:44217:44217)] 2024-11-24T19:45:27,491 DEBUG [32a03c886245:36267 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-11-24T19:45:27,492 DEBUG [32a03c886245:36267 {}] balancer.BalancerClusterState(204): Hosts are {32a03c886245=0} racks are {/default-rack=0} 2024-11-24T19:45:27,496 DEBUG [32a03c886245:36267 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-24T19:45:27,497 DEBUG [32a03c886245:36267 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-24T19:45:27,497 DEBUG [32a03c886245:36267 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-24T19:45:27,497 DEBUG [32a03c886245:36267 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-24T19:45:27,497 DEBUG [32a03c886245:36267 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-24T19:45:27,497 DEBUG [32a03c886245:36267 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-24T19:45:27,497 INFO [32a03c886245:36267 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-24T19:45:27,497 INFO [32a03c886245:36267 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-24T19:45:27,497 INFO [32a03c886245:36267 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-24T19:45:27,497 DEBUG [32a03c886245:36267 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-24T19:45:27,498 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=32a03c886245,44439,1732477526570 2024-11-24T19:45:27,502 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 32a03c886245,44439,1732477526570, state=OPENING 2024-11-24T19:45:27,527 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-11-24T19:45:27,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:27,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:27,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:27,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:27,541 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-24T19:45:27,541 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:27,541 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:27,541 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=32a03c886245,44439,1732477526570}] 2024-11-24T19:45:27,541 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:27,542 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:27,699 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-24T19:45:27,703 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60343, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-24T19:45:27,710 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-11-24T19:45:27,710 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-24T19:45:27,713 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=32a03c886245%2C44439%2C1732477526570.meta, suffix=.meta, logDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,44439,1732477526570, archiveDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/oldWALs, maxLogs=32 2024-11-24T19:45:27,715 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 32a03c886245%2C44439%2C1732477526570.meta.1732477527715.meta 2024-11-24T19:45:27,725 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/WALs/32a03c886245,44439,1732477526570/32a03c886245%2C44439%2C1732477526570.meta.1732477527715.meta 2024-11-24T19:45:27,730 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45169:45169),(127.0.0.1/127.0.0.1:44217:44217),(127.0.0.1/127.0.0.1:37763:37763)] 2024-11-24T19:45:27,731 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-11-24T19:45:27,731 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-11-24T19:45:27,731 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-11-24T19:45:27,732 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-11-24T19:45:27,732 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-11-24T19:45:27,732 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:27,732 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-11-24T19:45:27,732 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-11-24T19:45:27,734 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-24T19:45:27,735 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-24T19:45:27,735 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,736 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:27,736 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-24T19:45:27,737 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-24T19:45:27,737 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,737 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:27,737 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-24T19:45:27,738 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-24T19:45:27,738 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,739 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:27,739 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-24T19:45:27,740 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-24T19:45:27,740 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,740 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-24T19:45:27,741 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-24T19:45:27,741 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740 2024-11-24T19:45:27,743 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740 2024-11-24T19:45:27,744 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-24T19:45:27,744 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-24T19:45:27,745 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-24T19:45:27,746 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-24T19:45:27,747 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=75361121, jitterRate=0.12296821177005768}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-24T19:45:27,747 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-11-24T19:45:27,748 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1732477527732Writing region info on filesystem at 1732477527732Initializing all the Stores at 1732477527733 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477527733Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477527734 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477527734Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1732477527734Cleaning up temporary data from old regions at 1732477527744 (+10 ms)Running coprocessor post-open hooks at 1732477527747 (+3 ms)Region opened successfully at 1732477527748 (+1 ms) 2024-11-24T19:45:27,750 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1732477527698 2024-11-24T19:45:27,757 DEBUG [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-11-24T19:45:27,758 INFO [RS_OPEN_META-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-11-24T19:45:27,759 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=32a03c886245,44439,1732477526570 2024-11-24T19:45:27,760 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 32a03c886245,44439,1732477526570, state=OPEN 2024-11-24T19:45:27,802 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-24T19:45:27,802 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-24T19:45:27,802 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-24T19:45:27,802 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-24T19:45:27,802 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=32a03c886245,44439,1732477526570 2024-11-24T19:45:27,802 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:27,802 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:27,802 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:27,802 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-24T19:45:27,808 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-11-24T19:45:27,808 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=32a03c886245,44439,1732477526570 in 261 msec 2024-11-24T19:45:27,813 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-11-24T19:45:27,814 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 721 msec 2024-11-24T19:45:27,815 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-24T19:45:27,815 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-11-24T19:45:27,816 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-24T19:45:27,817 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=32a03c886245,44439,1732477526570, seqNum=-1] 2024-11-24T19:45:27,817 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-24T19:45:27,818 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48383, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-24T19:45:27,828 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 799 msec 2024-11-24T19:45:27,828 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1732477527828, completionTime=-1 2024-11-24T19:45:27,828 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-11-24T19:45:27,828 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-11-24T19:45:27,831 INFO [master/32a03c886245:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-11-24T19:45:27,831 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1732477587831 2024-11-24T19:45:27,831 INFO [master/32a03c886245:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1732477647831 2024-11-24T19:45:27,831 INFO [master/32a03c886245:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-11-24T19:45:27,831 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,36267,1732477526385-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,831 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,36267,1732477526385-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,832 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,36267,1732477526385-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,832 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-32a03c886245:36267, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,832 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,832 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,834 DEBUG [master/32a03c886245:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-11-24T19:45:27,836 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.131sec 2024-11-24T19:45:27,837 INFO [master/32a03c886245:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-11-24T19:45:27,837 INFO [master/32a03c886245:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-11-24T19:45:27,837 INFO [master/32a03c886245:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-11-24T19:45:27,837 INFO [master/32a03c886245:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-11-24T19:45:27,837 INFO [master/32a03c886245:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-11-24T19:45:27,837 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,36267,1732477526385-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-24T19:45:27,837 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,36267,1732477526385-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-11-24T19:45:27,840 DEBUG [master/32a03c886245:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-11-24T19:45:27,840 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-11-24T19:45:27,840 INFO [master/32a03c886245:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=32a03c886245,36267,1732477526385-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-24T19:45:27,869 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@79af53ce, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-24T19:45:27,869 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 32a03c886245,36267,-1 for getting cluster id 2024-11-24T19:45:27,869 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-11-24T19:45:27,870 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '416c9059-ca21-4b55-9b45-3d71930a7e06' 2024-11-24T19:45:27,871 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-11-24T19:45:27,871 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "416c9059-ca21-4b55-9b45-3d71930a7e06" 2024-11-24T19:45:27,871 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6feed621, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-24T19:45:27,871 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [32a03c886245,36267,-1] 2024-11-24T19:45:27,871 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-11-24T19:45:27,872 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:27,873 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33890, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-11-24T19:45:27,874 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@49ec70d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-24T19:45:27,875 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-24T19:45:27,876 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=32a03c886245,44439,1732477526570, seqNum=-1] 2024-11-24T19:45:27,876 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-24T19:45:27,878 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55700, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-24T19:45:27,880 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=32a03c886245,36267,1732477526385 2024-11-24T19:45:27,881 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-11-24T19:45:27,882 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is 32a03c886245,36267,1732477526385 2024-11-24T19:45:27,882 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@36210109 2024-11-24T19:45:27,883 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-11-24T19:45:27,885 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33894, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-11-24T19:45:27,886 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-24T19:45:27,888 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-11-24T19:45:27,890 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-11-24T19:45:27,890 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:27,890 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-11-24T19:45:27,891 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:27,892 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-24T19:45:27,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741837_1013 (size=392) 2024-11-24T19:45:27,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741837_1013 (size=392) 2024-11-24T19:45:27,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741837_1013 (size=392) 2024-11-24T19:45:27,907 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => bdcb6530204590ca7a1f1b7e22bee048, NAME => 'TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448 2024-11-24T19:45:27,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741838_1014 (size=51) 2024-11-24T19:45:27,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741838_1014 (size=51) 2024-11-24T19:45:27,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741838_1014 (size=51) 2024-11-24T19:45:27,920 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:27,920 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing bdcb6530204590ca7a1f1b7e22bee048, disabling compactions & flushes 2024-11-24T19:45:27,920 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:27,920 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:27,920 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. after waiting 0 ms 2024-11-24T19:45:27,921 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:27,921 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:27,921 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for bdcb6530204590ca7a1f1b7e22bee048: Waiting for close lock at 1732477527920Disabling compacts and flushes for region at 1732477527920Disabling writes for close at 1732477527920Writing region close event to WAL at 1732477527921 (+1 ms)Closed at 1732477527921 2024-11-24T19:45:27,922 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-11-24T19:45:27,923 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1732477527923"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1732477527923"}]},"ts":"1732477527923"} 2024-11-24T19:45:27,926 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-11-24T19:45:27,928 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-11-24T19:45:27,928 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1732477527928"}]},"ts":"1732477527928"} 2024-11-24T19:45:27,931 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-11-24T19:45:27,932 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {32a03c886245=0} racks are {/default-rack=0} 2024-11-24T19:45:27,933 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-24T19:45:27,933 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-24T19:45:27,933 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-24T19:45:27,933 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-24T19:45:27,933 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-24T19:45:27,933 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-24T19:45:27,933 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-24T19:45:27,933 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-24T19:45:27,933 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-24T19:45:27,933 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-24T19:45:27,933 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bdcb6530204590ca7a1f1b7e22bee048, ASSIGN}] 2024-11-24T19:45:27,936 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bdcb6530204590ca7a1f1b7e22bee048, ASSIGN 2024-11-24T19:45:27,937 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bdcb6530204590ca7a1f1b7e22bee048, ASSIGN; state=OFFLINE, location=32a03c886245,44439,1732477526570; forceNewPlan=false, retain=false 2024-11-24T19:45:28,003 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:28,088 INFO [32a03c886245:36267 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-11-24T19:45:28,088 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=bdcb6530204590ca7a1f1b7e22bee048, regionState=OPENING, regionLocation=32a03c886245,44439,1732477526570 2024-11-24T19:45:28,092 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bdcb6530204590ca7a1f1b7e22bee048, ASSIGN because future has completed 2024-11-24T19:45:28,092 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure bdcb6530204590ca7a1f1b7e22bee048, server=32a03c886245,44439,1732477526570}] 2024-11-24T19:45:28,212 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:28,253 INFO [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,254 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => bdcb6530204590ca7a1f1b7e22bee048, NAME => 'TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048.', STARTKEY => '', ENDKEY => ''} 2024-11-24T19:45:28,254 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,254 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-24T19:45:28,254 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,254 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,256 INFO [StoreOpener-bdcb6530204590ca7a1f1b7e22bee048-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,258 INFO [StoreOpener-bdcb6530204590ca7a1f1b7e22bee048-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region bdcb6530204590ca7a1f1b7e22bee048 columnFamilyName cf 2024-11-24T19:45:28,258 DEBUG [StoreOpener-bdcb6530204590ca7a1f1b7e22bee048-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-24T19:45:28,259 INFO [StoreOpener-bdcb6530204590ca7a1f1b7e22bee048-1 {}] regionserver.HStore(327): Store=bdcb6530204590ca7a1f1b7e22bee048/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-24T19:45:28,259 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,260 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,261 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,261 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,261 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,264 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,266 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-24T19:45:28,267 INFO [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened bdcb6530204590ca7a1f1b7e22bee048; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71393610, jitterRate=0.06384769082069397}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-11-24T19:45:28,267 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,268 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for bdcb6530204590ca7a1f1b7e22bee048: Running coprocessor pre-open hook at 1732477528255Writing region info on filesystem at 1732477528255Initializing all the Stores at 1732477528256 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1732477528256Cleaning up temporary data from old regions at 1732477528261 (+5 ms)Running coprocessor post-open hooks at 1732477528267 (+6 ms)Region opened successfully at 1732477528268 (+1 ms) 2024-11-24T19:45:28,269 INFO [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048., pid=6, masterSystemTime=1732477528246 2024-11-24T19:45:28,272 DEBUG [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,273 INFO [RS_OPEN_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,274 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=bdcb6530204590ca7a1f1b7e22bee048, regionState=OPEN, openSeqNum=2, regionLocation=32a03c886245,44439,1732477526570 2024-11-24T19:45:28,275 WARN [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36267 {}] assignment.AssignmentManager(1543): Unable to acquire lock for regionNode state=OPEN, location=32a03c886245,44439,1732477526570, table=TestHBaseWalOnEC, region=bdcb6530204590ca7a1f1b7e22bee048. It is likely that another thread is currently holding the lock. To avoid deadlock, skip execution for now. 2024-11-24T19:45:28,277 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure bdcb6530204590ca7a1f1b7e22bee048, server=32a03c886245,44439,1732477526570 because future has completed 2024-11-24T19:45:28,283 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-11-24T19:45:28,283 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure bdcb6530204590ca7a1f1b7e22bee048, server=32a03c886245,44439,1732477526570 in 187 msec 2024-11-24T19:45:28,288 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-11-24T19:45:28,288 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=bdcb6530204590ca7a1f1b7e22bee048, ASSIGN in 350 msec 2024-11-24T19:45:28,289 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-11-24T19:45:28,289 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1732477528289"}]},"ts":"1732477528289"} 2024-11-24T19:45:28,292 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-11-24T19:45:28,294 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-11-24T19:45:28,297 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 408 msec 2024-11-24T19:45:28,523 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-24T19:45:28,523 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-11-24T19:45:28,523 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-11-24T19:45:28,524 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-24T19:45:28,528 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-11-24T19:45:28,529 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-24T19:45:28,529 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-11-24T19:45:28,533 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048., hostname=32a03c886245,44439,1732477526570, seqNum=2] 2024-11-24T19:45:28,538 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-11-24T19:45:28,540 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-11-24T19:45:28,541 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-24T19:45:28,541 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-11-24T19:45:28,543 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-11-24T19:45:28,543 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-11-24T19:45:28,652 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-24T19:45:28,697 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=44439 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-11-24T19:45:28,698 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,698 INFO [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing bdcb6530204590ca7a1f1b7e22bee048 1/1 column families, dataSize=32 B heapSize=360 B 2024-11-24T19:45:28,715 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048/.tmp/cf/b229439b98c24ff68e9d7a0bc9fe10f1 is 36, key is row/cf:cq/1732477528535/Put/seqid=0 2024-11-24T19:45:28,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741839_1015 (size=4787) 2024-11-24T19:45:28,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741839_1015 (size=4787) 2024-11-24T19:45:28,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741839_1015 (size=4787) 2024-11-24T19:45:28,723 INFO [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048/.tmp/cf/b229439b98c24ff68e9d7a0bc9fe10f1 2024-11-24T19:45:28,732 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048/.tmp/cf/b229439b98c24ff68e9d7a0bc9fe10f1 as hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048/cf/b229439b98c24ff68e9d7a0bc9fe10f1 2024-11-24T19:45:28,741 INFO [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048/cf/b229439b98c24ff68e9d7a0bc9fe10f1, entries=1, sequenceid=5, filesize=4.7 K 2024-11-24T19:45:28,742 INFO [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for bdcb6530204590ca7a1f1b7e22bee048 in 44ms, sequenceid=5, compaction requested=false 2024-11-24T19:45:28,742 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for bdcb6530204590ca7a1f1b7e22bee048: 2024-11-24T19:45:28,742 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,743 DEBUG [RS_FLUSH_OPERATIONS-regionserver/32a03c886245:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-11-24T19:45:28,743 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-11-24T19:45:28,748 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-11-24T19:45:28,748 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 202 msec 2024-11-24T19:45:28,753 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 212 msec 2024-11-24T19:45:28,863 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36267 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-24T19:45:28,863 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-11-24T19:45:28,868 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-11-24T19:45:28,868 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-24T19:45:28,869 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:28,869 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:28,869 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:28,869 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-11-24T19:45:28,869 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-11-24T19:45:28,869 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=257125643, stopped=false 2024-11-24T19:45:28,869 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=32a03c886245,36267,1732477526385 2024-11-24T19:45:28,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:28,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:28,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:28,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:28,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-24T19:45:28,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:28,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:28,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:28,927 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:28,927 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-24T19:45:28,927 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:28,927 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-24T19:45:28,927 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:28,928 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:28,928 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:28,928 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-24T19:45:28,928 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32a03c886245,44439,1732477526570' ***** 2024-11-24T19:45:28,928 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-24T19:45:28,928 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32a03c886245,35081,1732477526607' ***** 2024-11-24T19:45:28,928 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-24T19:45:28,928 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '32a03c886245,45177,1732477526644' ***** 2024-11-24T19:45:28,928 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-24T19:45:28,928 INFO [RS:1;32a03c886245:35081 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-24T19:45:28,928 INFO [RS:1;32a03c886245:35081 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-24T19:45:28,929 INFO [RS:2;32a03c886245:45177 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-24T19:45:28,929 INFO [RS:1;32a03c886245:35081 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-24T19:45:28,929 INFO [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(959): stopping server 32a03c886245,35081,1732477526607 2024-11-24T19:45:28,929 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-24T19:45:28,929 INFO [RS:1;32a03c886245:35081 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-24T19:45:28,929 INFO [RS:1;32a03c886245:35081 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;32a03c886245:35081. 2024-11-24T19:45:28,929 DEBUG [RS:1;32a03c886245:35081 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:28,929 DEBUG [RS:1;32a03c886245:35081 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:28,929 INFO [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(976): stopping server 32a03c886245,35081,1732477526607; all regions closed. 2024-11-24T19:45:28,929 INFO [RS:0;32a03c886245:44439 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-24T19:45:28,929 INFO [RS:0;32a03c886245:44439 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-24T19:45:28,929 INFO [RS:0;32a03c886245:44439 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-24T19:45:28,929 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(3091): Received CLOSE for bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,929 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-24T19:45:28,930 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-24T19:45:28,930 INFO [RS:2;32a03c886245:45177 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-24T19:45:28,930 INFO [RS:2;32a03c886245:45177 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-24T19:45:28,930 INFO [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(959): stopping server 32a03c886245,45177,1732477526644 2024-11-24T19:45:28,930 INFO [RS:2;32a03c886245:45177 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-24T19:45:28,930 INFO [RS:2;32a03c886245:45177 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;32a03c886245:45177. 2024-11-24T19:45:28,930 DEBUG [RS:2;32a03c886245:45177 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:28,930 DEBUG [RS:2;32a03c886245:45177 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:28,930 INFO [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(976): stopping server 32a03c886245,45177,1732477526644; all regions closed. 2024-11-24T19:45:28,930 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,931 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,931 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,931 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,931 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,932 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(959): stopping server 32a03c886245,44439,1732477526570 2024-11-24T19:45:28,932 INFO [RS:0;32a03c886245:44439 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-24T19:45:28,932 INFO [RS:0;32a03c886245:44439 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;32a03c886245:44439. 2024-11-24T19:45:28,932 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing bdcb6530204590ca7a1f1b7e22bee048, disabling compactions & flushes 2024-11-24T19:45:28,932 INFO [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,932 DEBUG [RS:0;32a03c886245:44439 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-24T19:45:28,932 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,932 DEBUG [RS:0;32a03c886245:44439 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:28,932 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. after waiting 0 ms 2024-11-24T19:45:28,933 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,933 INFO [RS:0;32a03c886245:44439 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-24T19:45:28,933 INFO [RS:0;32a03c886245:44439 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-24T19:45:28,933 INFO [RS:0;32a03c886245:44439 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-24T19:45:28,933 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-11-24T19:45:28,933 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,933 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,933 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,933 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,934 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:28,936 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741833_1009 (size=93) 2024-11-24T19:45:28,936 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741833_1009 (size=93) 2024-11-24T19:45:28,938 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-11-24T19:45:28,939 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1325): Online Regions={bdcb6530204590ca7a1f1b7e22bee048=TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048., 1588230740=hbase:meta,,1.1588230740} 2024-11-24T19:45:28,939 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, bdcb6530204590ca7a1f1b7e22bee048 2024-11-24T19:45:28,939 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-24T19:45:28,939 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-24T19:45:28,939 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-24T19:45:28,939 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-24T19:45:28,939 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-24T19:45:28,939 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-11-24T19:45:28,942 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741833_1009 (size=93) 2024-11-24T19:45:28,944 DEBUG [RS:1;32a03c886245:35081 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/oldWALs 2024-11-24T19:45:28,945 INFO [RS:1;32a03c886245:35081 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 32a03c886245%2C35081%2C1732477526607:(num 1732477527311) 2024-11-24T19:45:28,945 DEBUG [RS:1;32a03c886245:35081 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:28,945 INFO [RS:1;32a03c886245:35081 {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:28,945 INFO [RS:1;32a03c886245:35081 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-24T19:45:28,945 INFO [RS:1;32a03c886245:35081 {}] hbase.ChoreService(370): Chore service for: regionserver/32a03c886245:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-11-24T19:45:28,946 INFO [regionserver/32a03c886245:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-24T19:45:28,946 INFO [RS:1;32a03c886245:35081 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-24T19:45:28,947 INFO [RS:1;32a03c886245:35081 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-24T19:45:28,947 INFO [RS:1;32a03c886245:35081 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-24T19:45:28,947 INFO [RS:1;32a03c886245:35081 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-24T19:45:28,947 INFO [RS:1;32a03c886245:35081 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:35081 2024-11-24T19:45:28,948 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741834_1010 (size=93) 2024-11-24T19:45:28,948 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/default/TestHBaseWalOnEC/bdcb6530204590ca7a1f1b7e22bee048/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-11-24T19:45:28,949 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741834_1010 (size=93) 2024-11-24T19:45:28,949 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741834_1010 (size=93) 2024-11-24T19:45:28,950 INFO [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,950 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for bdcb6530204590ca7a1f1b7e22bee048: Waiting for close lock at 1732477528932Running coprocessor pre-close hooks at 1732477528932Disabling compacts and flushes for region at 1732477528932Disabling writes for close at 1732477528933 (+1 ms)Writing region close event to WAL at 1732477528933Running coprocessor post-close hooks at 1732477528950 (+17 ms)Closed at 1732477528950 2024-11-24T19:45:28,950 DEBUG [RS_CLOSE_REGION-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048. 2024-11-24T19:45:28,953 DEBUG [RS:2;32a03c886245:45177 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/oldWALs 2024-11-24T19:45:28,953 INFO [RS:2;32a03c886245:45177 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 32a03c886245%2C45177%2C1732477526644:(num 1732477527312) 2024-11-24T19:45:28,954 DEBUG [RS:2;32a03c886245:45177 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:28,954 INFO [RS:2;32a03c886245:45177 {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:28,954 INFO [RS:2;32a03c886245:45177 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-24T19:45:28,954 INFO [RS:2;32a03c886245:45177 {}] hbase.ChoreService(370): Chore service for: regionserver/32a03c886245:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-24T19:45:28,954 INFO [RS:2;32a03c886245:45177 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-24T19:45:28,954 INFO [RS:2;32a03c886245:45177 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-24T19:45:28,954 INFO [RS:2;32a03c886245:45177 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-24T19:45:28,954 INFO [RS:2;32a03c886245:45177 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-24T19:45:28,955 INFO [RS:2;32a03c886245:45177 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:45177 2024-11-24T19:45:28,955 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32a03c886245,35081,1732477526607 2024-11-24T19:45:28,955 INFO [RS:1;32a03c886245:35081 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-24T19:45:28,955 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-24T19:45:28,955 INFO [regionserver/32a03c886245:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-24T19:45:28,970 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/info/4c0b8a76d7f24efda9c429fe90427ed9 is 153, key is TestHBaseWalOnEC,,1732477527885.bdcb6530204590ca7a1f1b7e22bee048./info:regioninfo/1732477528274/Put/seqid=0 2024-11-24T19:45:28,971 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32a03c886245,45177,1732477526644 2024-11-24T19:45:28,971 INFO [RS:2;32a03c886245:45177 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-24T19:45:28,972 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32a03c886245,45177,1732477526644] 2024-11-24T19:45:28,975 INFO [regionserver/32a03c886245:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:28,975 INFO [regionserver/32a03c886245:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:28,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741840_1016 (size=6637) 2024-11-24T19:45:28,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741840_1016 (size=6637) 2024-11-24T19:45:28,979 INFO [regionserver/32a03c886245:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:28,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741840_1016 (size=6637) 2024-11-24T19:45:28,980 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32a03c886245,45177,1732477526644 already deleted, retry=false 2024-11-24T19:45:28,980 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32a03c886245,45177,1732477526644 expired; onlineServers=2 2024-11-24T19:45:28,980 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32a03c886245,35081,1732477526607] 2024-11-24T19:45:28,980 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/info/4c0b8a76d7f24efda9c429fe90427ed9 2024-11-24T19:45:28,988 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32a03c886245,35081,1732477526607 already deleted, retry=false 2024-11-24T19:45:28,988 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32a03c886245,35081,1732477526607 expired; onlineServers=1 2024-11-24T19:45:29,007 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/ns/de11801fac5f4a5e969023f2ebe27143 is 43, key is default/ns:d/1732477527819/Put/seqid=0 2024-11-24T19:45:29,009 WARN [IPC Server handler 2 on default port 42003 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:29,009 WARN [IPC Server handler 2 on default port 42003 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:29,009 WARN [IPC Server handler 2 on default port 42003 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:29,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741841_1017 (size=5153) 2024-11-24T19:45:29,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741841_1017 (size=5153) 2024-11-24T19:45:29,019 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/ns/de11801fac5f4a5e969023f2ebe27143 2024-11-24T19:45:29,056 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/table/7a3f943f86ce4c55a13da360a1fc13f7 is 52, key is TestHBaseWalOnEC/table:state/1732477528289/Put/seqid=0 2024-11-24T19:45:29,058 WARN [IPC Server handler 1 on default port 42003 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-24T19:45:29,058 WARN [IPC Server handler 1 on default port 42003 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-24T19:45:29,058 WARN [IPC Server handler 1 on default port 42003 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-24T19:45:29,064 INFO [RS:1;32a03c886245:35081 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-24T19:45:29,064 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:29,064 INFO [RS:1;32a03c886245:35081 {}] regionserver.HRegionServer(1031): Exiting; stopping=32a03c886245,35081,1732477526607; zookeeper connection closed. 2024-11-24T19:45:29,064 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35081-0x1016faab4230002, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:29,071 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@53484c39 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@53484c39 2024-11-24T19:45:29,077 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741842_1018 (size=5249) 2024-11-24T19:45:29,078 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741842_1018 (size=5249) 2024-11-24T19:45:29,079 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/table/7a3f943f86ce4c55a13da360a1fc13f7 2024-11-24T19:45:29,080 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:29,080 INFO [RS:2;32a03c886245:45177 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-24T19:45:29,080 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45177-0x1016faab4230003, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:29,080 INFO [RS:2;32a03c886245:45177 {}] regionserver.HRegionServer(1031): Exiting; stopping=32a03c886245,45177,1732477526644; zookeeper connection closed. 2024-11-24T19:45:29,080 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@1b89a012 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@1b89a012 2024-11-24T19:45:29,090 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/info/4c0b8a76d7f24efda9c429fe90427ed9 as hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/info/4c0b8a76d7f24efda9c429fe90427ed9 2024-11-24T19:45:29,102 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/info/4c0b8a76d7f24efda9c429fe90427ed9, entries=10, sequenceid=11, filesize=6.5 K 2024-11-24T19:45:29,105 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/ns/de11801fac5f4a5e969023f2ebe27143 as hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/ns/de11801fac5f4a5e969023f2ebe27143 2024-11-24T19:45:29,121 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/ns/de11801fac5f4a5e969023f2ebe27143, entries=2, sequenceid=11, filesize=5.0 K 2024-11-24T19:45:29,123 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/.tmp/table/7a3f943f86ce4c55a13da360a1fc13f7 as hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/table/7a3f943f86ce4c55a13da360a1fc13f7 2024-11-24T19:45:29,132 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/table/7a3f943f86ce4c55a13da360a1fc13f7, entries=2, sequenceid=11, filesize=5.1 K 2024-11-24T19:45:29,134 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 195ms, sequenceid=11, compaction requested=false 2024-11-24T19:45:29,139 DEBUG [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-11-24T19:45:29,155 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-11-24T19:45:29,156 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-24T19:45:29,156 INFO [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-24T19:45:29,157 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1732477528939Running coprocessor pre-close hooks at 1732477528939Disabling compacts and flushes for region at 1732477528939Disabling writes for close at 1732477528939Obtaining lock to block concurrent updates at 1732477528939Preparing flush snapshotting stores in 1588230740 at 1732477528939Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1732477528940 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1732477528941 (+1 ms)Flushing 1588230740/info: creating writer at 1732477528941Flushing 1588230740/info: appending metadata at 1732477528969 (+28 ms)Flushing 1588230740/info: closing flushed file at 1732477528969Flushing 1588230740/ns: creating writer at 1732477528990 (+21 ms)Flushing 1588230740/ns: appending metadata at 1732477529007 (+17 ms)Flushing 1588230740/ns: closing flushed file at 1732477529007Flushing 1588230740/table: creating writer at 1732477529029 (+22 ms)Flushing 1588230740/table: appending metadata at 1732477529056 (+27 ms)Flushing 1588230740/table: closing flushed file at 1732477529056Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7a215ced: reopening flushed file at 1732477529088 (+32 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@40ebf11d: reopening flushed file at 1732477529103 (+15 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7abd9197: reopening flushed file at 1732477529121 (+18 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 195ms, sequenceid=11, compaction requested=false at 1732477529134 (+13 ms)Writing region close event to WAL at 1732477529150 (+16 ms)Running coprocessor post-close hooks at 1732477529156 (+6 ms)Closed at 1732477529156 2024-11-24T19:45:29,157 DEBUG [RS_CLOSE_META-regionserver/32a03c886245:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-11-24T19:45:29,180 INFO [regionserver/32a03c886245:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-11-24T19:45:29,180 INFO [regionserver/32a03c886245:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-11-24T19:45:29,339 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(976): stopping server 32a03c886245,44439,1732477526570; all regions closed. 2024-11-24T19:45:29,340 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,340 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,340 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,340 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,340 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741836_1012 (size=2751) 2024-11-24T19:45:29,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741836_1012 (size=2751) 2024-11-24T19:45:29,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741836_1012 (size=2751) 2024-11-24T19:45:29,345 DEBUG [RS:0;32a03c886245:44439 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/oldWALs 2024-11-24T19:45:29,346 INFO [RS:0;32a03c886245:44439 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 32a03c886245%2C44439%2C1732477526570.meta:.meta(num 1732477527715) 2024-11-24T19:45:29,346 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,346 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,346 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,346 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,346 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,349 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741835_1011 (size=1298) 2024-11-24T19:45:29,349 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741835_1011 (size=1298) 2024-11-24T19:45:29,349 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741835_1011 (size=1298) 2024-11-24T19:45:29,352 DEBUG [RS:0;32a03c886245:44439 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/oldWALs 2024-11-24T19:45:29,352 INFO [RS:0;32a03c886245:44439 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 32a03c886245%2C44439%2C1732477526570:(num 1732477527313) 2024-11-24T19:45:29,352 DEBUG [RS:0;32a03c886245:44439 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-24T19:45:29,352 INFO [RS:0;32a03c886245:44439 {}] regionserver.LeaseManager(133): Closed leases 2024-11-24T19:45:29,352 INFO [RS:0;32a03c886245:44439 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-24T19:45:29,352 INFO [RS:0;32a03c886245:44439 {}] hbase.ChoreService(370): Chore service for: regionserver/32a03c886245:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-24T19:45:29,352 INFO [RS:0;32a03c886245:44439 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-24T19:45:29,352 INFO [regionserver/32a03c886245:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-24T19:45:29,353 INFO [RS:0;32a03c886245:44439 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:44439 2024-11-24T19:45:29,385 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-24T19:45:29,385 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/32a03c886245,44439,1732477526570 2024-11-24T19:45:29,385 INFO [RS:0;32a03c886245:44439 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-24T19:45:29,385 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [32a03c886245,44439,1732477526570] 2024-11-24T19:45:29,401 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/32a03c886245,44439,1732477526570 already deleted, retry=false 2024-11-24T19:45:29,401 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 32a03c886245,44439,1732477526570 expired; onlineServers=0 2024-11-24T19:45:29,401 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '32a03c886245,36267,1732477526385' ***** 2024-11-24T19:45:29,401 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-11-24T19:45:29,401 INFO [M:0;32a03c886245:36267 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-24T19:45:29,401 INFO [M:0;32a03c886245:36267 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-24T19:45:29,402 DEBUG [M:0;32a03c886245:36267 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-11-24T19:45:29,402 DEBUG [M:0;32a03c886245:36267 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-11-24T19:45:29,402 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-11-24T19:45:29,402 DEBUG [master/32a03c886245:0:becomeActiveMaster-HFileCleaner.large.0-1732477527035 {}] cleaner.HFileCleaner(306): Exit Thread[master/32a03c886245:0:becomeActiveMaster-HFileCleaner.large.0-1732477527035,5,FailOnTimeoutGroup] 2024-11-24T19:45:29,402 DEBUG [master/32a03c886245:0:becomeActiveMaster-HFileCleaner.small.0-1732477527035 {}] cleaner.HFileCleaner(306): Exit Thread[master/32a03c886245:0:becomeActiveMaster-HFileCleaner.small.0-1732477527035,5,FailOnTimeoutGroup] 2024-11-24T19:45:29,402 INFO [M:0;32a03c886245:36267 {}] hbase.ChoreService(370): Chore service for: master/32a03c886245:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-11-24T19:45:29,402 INFO [M:0;32a03c886245:36267 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-24T19:45:29,403 DEBUG [M:0;32a03c886245:36267 {}] master.HMaster(1795): Stopping service threads 2024-11-24T19:45:29,403 INFO [M:0;32a03c886245:36267 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-11-24T19:45:29,403 INFO [M:0;32a03c886245:36267 {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-24T19:45:29,403 INFO [M:0;32a03c886245:36267 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-11-24T19:45:29,403 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-11-24T19:45:29,413 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-11-24T19:45:29,413 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-24T19:45:29,413 DEBUG [M:0;32a03c886245:36267 {}] zookeeper.ZKUtil(347): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-11-24T19:45:29,413 WARN [M:0;32a03c886245:36267 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-11-24T19:45:29,414 INFO [M:0;32a03c886245:36267 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/.lastflushedseqids 2024-11-24T19:45:29,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741843_1019 (size=127) 2024-11-24T19:45:29,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741843_1019 (size=127) 2024-11-24T19:45:29,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741843_1019 (size=127) 2024-11-24T19:45:29,427 INFO [M:0;32a03c886245:36267 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-11-24T19:45:29,428 INFO [M:0;32a03c886245:36267 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-11-24T19:45:29,428 DEBUG [M:0;32a03c886245:36267 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-24T19:45:29,428 INFO [M:0;32a03c886245:36267 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:29,428 DEBUG [M:0;32a03c886245:36267 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:29,428 DEBUG [M:0;32a03c886245:36267 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-24T19:45:29,428 DEBUG [M:0;32a03c886245:36267 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:29,428 INFO [M:0;32a03c886245:36267 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.81 KB heapSize=34.10 KB 2024-11-24T19:45:29,448 DEBUG [M:0;32a03c886245:36267 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/737e2e5412ce4c09beefd915b3463f1a is 82, key is hbase:meta,,1/info:regioninfo/1732477527759/Put/seqid=0 2024-11-24T19:45:29,455 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741844_1020 (size=5672) 2024-11-24T19:45:29,455 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741844_1020 (size=5672) 2024-11-24T19:45:29,455 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741844_1020 (size=5672) 2024-11-24T19:45:29,456 INFO [M:0;32a03c886245:36267 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/737e2e5412ce4c09beefd915b3463f1a 2024-11-24T19:45:29,478 DEBUG [M:0;32a03c886245:36267 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/b59ff776997549b7952abd9bef812d03 is 747, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1732477528296/Put/seqid=0 2024-11-24T19:45:29,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741845_1021 (size=6437) 2024-11-24T19:45:29,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741845_1021 (size=6437) 2024-11-24T19:45:29,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741845_1021 (size=6437) 2024-11-24T19:45:29,486 INFO [M:0;32a03c886245:36267 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.12 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/b59ff776997549b7952abd9bef812d03 2024-11-24T19:45:29,493 INFO [RS:0;32a03c886245:44439 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-24T19:45:29,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:29,493 INFO [RS:0;32a03c886245:44439 {}] regionserver.HRegionServer(1031): Exiting; stopping=32a03c886245,44439,1732477526570; zookeeper connection closed. 2024-11-24T19:45:29,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44439-0x1016faab4230001, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:29,493 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@3cc2fa0e {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@3cc2fa0e 2024-11-24T19:45:29,494 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-11-24T19:45:29,507 DEBUG [M:0;32a03c886245:36267 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1ba8a7fa556e4403b160cbe3b4edc444 is 69, key is 32a03c886245,35081,1732477526607/rs:state/1732477527107/Put/seqid=0 2024-11-24T19:45:29,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741846_1022 (size=5294) 2024-11-24T19:45:29,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741846_1022 (size=5294) 2024-11-24T19:45:29,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741846_1022 (size=5294) 2024-11-24T19:45:29,515 INFO [M:0;32a03c886245:36267 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1ba8a7fa556e4403b160cbe3b4edc444 2024-11-24T19:45:29,522 DEBUG [M:0;32a03c886245:36267 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/737e2e5412ce4c09beefd915b3463f1a as hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/737e2e5412ce4c09beefd915b3463f1a 2024-11-24T19:45:29,529 INFO [M:0;32a03c886245:36267 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/737e2e5412ce4c09beefd915b3463f1a, entries=8, sequenceid=72, filesize=5.5 K 2024-11-24T19:45:29,530 DEBUG [M:0;32a03c886245:36267 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/b59ff776997549b7952abd9bef812d03 as hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/b59ff776997549b7952abd9bef812d03 2024-11-24T19:45:29,537 INFO [M:0;32a03c886245:36267 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/b59ff776997549b7952abd9bef812d03, entries=8, sequenceid=72, filesize=6.3 K 2024-11-24T19:45:29,538 DEBUG [M:0;32a03c886245:36267 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1ba8a7fa556e4403b160cbe3b4edc444 as hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/1ba8a7fa556e4403b160cbe3b4edc444 2024-11-24T19:45:29,546 INFO [M:0;32a03c886245:36267 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42003/user/jenkins/test-data/7c81cbc8-3e4d-af5b-bcd0-259e14c26448/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/1ba8a7fa556e4403b160cbe3b4edc444, entries=3, sequenceid=72, filesize=5.2 K 2024-11-24T19:45:29,548 INFO [M:0;32a03c886245:36267 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.81 KB/27450, heapSize ~33.80 KB/34616, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 120ms, sequenceid=72, compaction requested=false 2024-11-24T19:45:29,549 INFO [M:0;32a03c886245:36267 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-24T19:45:29,550 DEBUG [M:0;32a03c886245:36267 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1732477529428Disabling compacts and flushes for region at 1732477529428Disabling writes for close at 1732477529428Obtaining lock to block concurrent updates at 1732477529428Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1732477529428Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27450, getHeapSize=34856, getOffHeapSize=0, getCellsCount=85 at 1732477529429 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1732477529430 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1732477529431 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1732477529448 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1732477529448Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1732477529462 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1732477529478 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1732477529478Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1732477529491 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1732477529506 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1732477529506Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@63d399e2: reopening flushed file at 1732477529520 (+14 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@66aaf52a: reopening flushed file at 1732477529529 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@19085854: reopening flushed file at 1732477529537 (+8 ms)Finished flush of dataSize ~26.81 KB/27450, heapSize ~33.80 KB/34616, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 120ms, sequenceid=72, compaction requested=false at 1732477529548 (+11 ms)Writing region close event to WAL at 1732477529549 (+1 ms)Closed at 1732477529549 2024-11-24T19:45:29,550 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,550 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,550 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,550 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,550 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-24T19:45:29,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36477 is added to blk_1073741830_1006 (size=32653) 2024-11-24T19:45:29,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34183 is added to blk_1073741830_1006 (size=32653) 2024-11-24T19:45:29,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36373 is added to blk_1073741830_1006 (size=32653) 2024-11-24T19:45:29,553 INFO [M:0;32a03c886245:36267 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-11-24T19:45:29,553 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-24T19:45:29,554 INFO [M:0;32a03c886245:36267 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36267 2024-11-24T19:45:29,554 INFO [M:0;32a03c886245:36267 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-24T19:45:29,680 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:29,680 INFO [M:0;32a03c886245:36267 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-24T19:45:29,680 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36267-0x1016faab4230000, quorum=127.0.0.1:55315, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-24T19:45:29,684 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@51eb0253{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:29,685 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@38066021{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-24T19:45:29,685 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-24T19:45:29,686 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@498db938{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-24T19:45:29,686 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@25de4fb7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir/,STOPPED} 2024-11-24T19:45:29,687 WARN [BP-41452304-172.17.0.2-1732477524022 heartbeating to localhost/127.0.0.1:42003 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-24T19:45:29,687 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-24T19:45:29,688 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-24T19:45:29,688 WARN [BP-41452304-172.17.0.2-1732477524022 heartbeating to localhost/127.0.0.1:42003 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-41452304-172.17.0.2-1732477524022 (Datanode Uuid de0a7a9d-20f5-40b1-a5e6-85b077b5d204) service to localhost/127.0.0.1:42003 2024-11-24T19:45:29,688 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data5/current/BP-41452304-172.17.0.2-1732477524022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:29,688 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data6/current/BP-41452304-172.17.0.2-1732477524022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:29,689 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-24T19:45:29,695 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@569dbdc8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:29,695 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@18c4cb78{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-24T19:45:29,695 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-24T19:45:29,695 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2b2d9a02{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-24T19:45:29,695 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@17cd1c9f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir/,STOPPED} 2024-11-24T19:45:29,697 WARN [BP-41452304-172.17.0.2-1732477524022 heartbeating to localhost/127.0.0.1:42003 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-24T19:45:29,697 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-24T19:45:29,697 WARN [BP-41452304-172.17.0.2-1732477524022 heartbeating to localhost/127.0.0.1:42003 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-41452304-172.17.0.2-1732477524022 (Datanode Uuid fd9037f1-a98d-4491-9463-e5df8ca942c3) service to localhost/127.0.0.1:42003 2024-11-24T19:45:29,697 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-24T19:45:29,697 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data3/current/BP-41452304-172.17.0.2-1732477524022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:29,697 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data4/current/BP-41452304-172.17.0.2-1732477524022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:29,698 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-24T19:45:29,699 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@484ab650{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-24T19:45:29,700 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1d09d7d5{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-24T19:45:29,700 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-24T19:45:29,700 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@d63bfae{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-24T19:45:29,700 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6c411bc1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir/,STOPPED} 2024-11-24T19:45:29,701 WARN [BP-41452304-172.17.0.2-1732477524022 heartbeating to localhost/127.0.0.1:42003 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-24T19:45:29,701 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-24T19:45:29,701 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-24T19:45:29,701 WARN [BP-41452304-172.17.0.2-1732477524022 heartbeating to localhost/127.0.0.1:42003 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-41452304-172.17.0.2-1732477524022 (Datanode Uuid 854a62d0-6f39-4749-a4be-2b3ae7e2c987) service to localhost/127.0.0.1:42003 2024-11-24T19:45:29,702 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data1/current/BP-41452304-172.17.0.2-1732477524022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:29,702 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/cluster_50952b5a-f46b-29d9-a2db-d9bf6252811c/data/data2/current/BP-41452304-172.17.0.2-1732477524022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-24T19:45:29,702 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-24T19:45:29,707 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5609bdf8{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-24T19:45:29,707 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7cbf28d2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-24T19:45:29,708 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-24T19:45:29,708 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5574eaf2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-24T19:45:29,708 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4c686d6e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/bbd6f963-c756-7591-ff70-781ed7102ee3/hadoop.log.dir/,STOPPED} 2024-11-24T19:45:29,715 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-11-24T19:45:29,745 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-11-24T19:45:29,751 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=146 (was 87) - Thread LEAK? -, OpenFileDescriptor=518 (was 433) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=336 (was 330) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=10618 (was 10813)