2024-12-06 07:53:02,777 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-06 07:53:02,789 main DEBUG Took 0.009808 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-06 07:53:02,789 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-06 07:53:02,790 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-06 07:53:02,791 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-06 07:53:02,792 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,801 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-06 07:53:02,819 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,821 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,822 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,822 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,822 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,823 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,824 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,824 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,824 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,825 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,825 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,826 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,826 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,826 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,827 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,827 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,828 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,828 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,828 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,829 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,829 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,829 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,830 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,830 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 07:53:02,830 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,831 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-06 07:53:02,832 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 07:53:02,834 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-06 07:53:02,835 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-06 07:53:02,836 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-06 07:53:02,837 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-06 07:53:02,837 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-06 07:53:02,846 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-06 07:53:02,849 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-06 07:53:02,850 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-06 07:53:02,851 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-06 07:53:02,851 main DEBUG createAppenders(={Console}) 2024-12-06 07:53:02,852 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-06 07:53:02,852 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-06 07:53:02,852 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-06 07:53:02,853 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-06 07:53:02,853 main DEBUG OutputStream closed 2024-12-06 07:53:02,853 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-06 07:53:02,854 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-06 07:53:02,854 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-06 07:53:02,931 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-06 07:53:02,934 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-06 07:53:02,935 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-06 07:53:02,936 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-06 07:53:02,936 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-06 07:53:02,937 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-06 07:53:02,937 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-06 07:53:02,937 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-06 07:53:02,937 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-06 07:53:02,938 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-06 07:53:02,938 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-06 07:53:02,938 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-06 07:53:02,939 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-06 07:53:02,939 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-06 07:53:02,939 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-06 07:53:02,939 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-06 07:53:02,940 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-06 07:53:02,940 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-06 07:53:02,943 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-06 07:53:02,943 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-06 07:53:02,943 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-06 07:53:02,944 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-06T07:53:02,960 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-06 07:53:02,963 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-06 07:53:02,963 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-06T07:53:03,248 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90 2024-12-06T07:53:03,273 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56, deleteOnExit=true 2024-12-06T07:53:03,274 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/test.cache.data in system properties and HBase conf 2024-12-06T07:53:03,274 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.tmp.dir in system properties and HBase conf 2024-12-06T07:53:03,275 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir in system properties and HBase conf 2024-12-06T07:53:03,276 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-06T07:53:03,276 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-06T07:53:03,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-06T07:53:03,389 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-06T07:53:03,484 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-06T07:53:03,487 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-06T07:53:03,488 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-06T07:53:03,488 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-06T07:53:03,489 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-06T07:53:03,489 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-06T07:53:03,490 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-06T07:53:03,490 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-06T07:53:03,491 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-06T07:53:03,491 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-06T07:53:03,491 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/nfs.dump.dir in system properties and HBase conf 2024-12-06T07:53:03,492 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/java.io.tmpdir in system properties and HBase conf 2024-12-06T07:53:03,492 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-06T07:53:03,492 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-06T07:53:03,493 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-06T07:53:04,291 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-06T07:53:04,368 INFO [Time-limited test {}] log.Log(170): Logging initialized @2295ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-06T07:53:04,447 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:04,512 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T07:53:04,534 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T07:53:04,534 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T07:53:04,535 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T07:53:04,549 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:04,552 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir/,AVAILABLE} 2024-12-06T07:53:04,553 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T07:53:04,744 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@76e4c45c{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/java.io.tmpdir/jetty-localhost-39119-hadoop-hdfs-3_4_1-tests_jar-_-any-1828004891360618754/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-06T07:53:04,751 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:39119} 2024-12-06T07:53:04,752 INFO [Time-limited test {}] server.Server(415): Started @2680ms 2024-12-06T07:53:05,126 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:05,132 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T07:53:05,133 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T07:53:05,133 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T07:53:05,133 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T07:53:05,134 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@550154bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir/,AVAILABLE} 2024-12-06T07:53:05,135 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1a2478ad{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T07:53:05,252 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4839957b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/java.io.tmpdir/jetty-localhost-44589-hadoop-hdfs-3_4_1-tests_jar-_-any-5876739129775987020/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:05,253 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5306f615{HTTP/1.1, (http/1.1)}{localhost:44589} 2024-12-06T07:53:05,253 INFO [Time-limited test {}] server.Server(415): Started @3181ms 2024-12-06T07:53:05,306 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T07:53:05,421 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:05,428 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T07:53:05,432 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T07:53:05,433 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T07:53:05,433 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T07:53:05,434 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6463ad04{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir/,AVAILABLE} 2024-12-06T07:53:05,435 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7fa8fa5c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T07:53:05,587 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1c6b8f01{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/java.io.tmpdir/jetty-localhost-32893-hadoop-hdfs-3_4_1-tests_jar-_-any-8639986527565442052/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:05,588 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11f28dd2{HTTP/1.1, (http/1.1)}{localhost:32893} 2024-12-06T07:53:05,588 INFO [Time-limited test {}] server.Server(415): Started @3516ms 2024-12-06T07:53:05,591 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T07:53:05,633 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:05,638 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T07:53:05,639 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T07:53:05,639 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T07:53:05,640 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-06T07:53:05,647 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@c62369b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir/,AVAILABLE} 2024-12-06T07:53:05,648 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@24f92c39{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T07:53:05,744 WARN [Thread-107 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data2/current/BP-1887890605-172.17.0.2-1733471584055/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:05,744 WARN [Thread-108 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data4/current/BP-1887890605-172.17.0.2-1733471584055/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:05,744 WARN [Thread-106 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data3/current/BP-1887890605-172.17.0.2-1733471584055/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:05,744 WARN [Thread-105 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data1/current/BP-1887890605-172.17.0.2-1733471584055/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:05,789 WARN [Thread-82 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T07:53:05,789 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T07:53:05,806 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2e59159d{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/java.io.tmpdir/jetty-localhost-35027-hadoop-hdfs-3_4_1-tests_jar-_-any-2850821560914356127/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:05,807 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@a8e922f{HTTP/1.1, (http/1.1)}{localhost:35027} 2024-12-06T07:53:05,807 INFO [Time-limited test {}] server.Server(415): Started @3735ms 2024-12-06T07:53:05,811 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T07:53:05,859 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xec49f4ff4586834f with lease ID 0xe61d5b3456557506: Processing first storage report for DS-2d469ba5-aa45-4caf-b546-a241f2464343 from datanode DatanodeRegistration(127.0.0.1:42163, datanodeUuid=b0aa09a7-b7db-4430-af7a-e791adc4248c, infoPort=44973, infoSecurePort=0, ipcPort=41073, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055) 2024-12-06T07:53:05,860 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xec49f4ff4586834f with lease ID 0xe61d5b3456557506: from storage DS-2d469ba5-aa45-4caf-b546-a241f2464343 node DatanodeRegistration(127.0.0.1:42163, datanodeUuid=b0aa09a7-b7db-4430-af7a-e791adc4248c, infoPort=44973, infoSecurePort=0, ipcPort=41073, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-06T07:53:05,861 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3d6fcc6da091c33f with lease ID 0xe61d5b3456557505: Processing first storage report for DS-45455d47-b25c-49ab-922e-4c174df95d0f from datanode DatanodeRegistration(127.0.0.1:45919, datanodeUuid=6f74a450-2703-4638-8c9d-1cfd0de2cab4, infoPort=36261, infoSecurePort=0, ipcPort=36887, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055) 2024-12-06T07:53:05,861 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3d6fcc6da091c33f with lease ID 0xe61d5b3456557505: from storage DS-45455d47-b25c-49ab-922e-4c174df95d0f node DatanodeRegistration(127.0.0.1:45919, datanodeUuid=6f74a450-2703-4638-8c9d-1cfd0de2cab4, infoPort=36261, infoSecurePort=0, ipcPort=36887, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T07:53:05,861 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xec49f4ff4586834f with lease ID 0xe61d5b3456557506: Processing first storage report for DS-b12ab067-068f-4855-968a-ac5159cb46a7 from datanode DatanodeRegistration(127.0.0.1:42163, datanodeUuid=b0aa09a7-b7db-4430-af7a-e791adc4248c, infoPort=44973, infoSecurePort=0, ipcPort=41073, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055) 2024-12-06T07:53:05,861 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xec49f4ff4586834f with lease ID 0xe61d5b3456557506: from storage DS-b12ab067-068f-4855-968a-ac5159cb46a7 node DatanodeRegistration(127.0.0.1:42163, datanodeUuid=b0aa09a7-b7db-4430-af7a-e791adc4248c, infoPort=44973, infoSecurePort=0, ipcPort=41073, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T07:53:05,862 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3d6fcc6da091c33f with lease ID 0xe61d5b3456557505: Processing first storage report for DS-e5223886-f567-4cf6-9340-4c8a121954f7 from datanode DatanodeRegistration(127.0.0.1:45919, datanodeUuid=6f74a450-2703-4638-8c9d-1cfd0de2cab4, infoPort=36261, infoSecurePort=0, ipcPort=36887, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055) 2024-12-06T07:53:05,862 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3d6fcc6da091c33f with lease ID 0xe61d5b3456557505: from storage DS-e5223886-f567-4cf6-9340-4c8a121954f7 node DatanodeRegistration(127.0.0.1:45919, datanodeUuid=6f74a450-2703-4638-8c9d-1cfd0de2cab4, infoPort=36261, infoSecurePort=0, ipcPort=36887, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-06T07:53:05,948 WARN [Thread-139 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data5/current/BP-1887890605-172.17.0.2-1733471584055/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:05,948 WARN [Thread-140 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data6/current/BP-1887890605-172.17.0.2-1733471584055/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:05,976 WARN [Thread-129 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T07:53:05,981 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc776167b283c2e89 with lease ID 0xe61d5b3456557507: Processing first storage report for DS-400f7f99-b834-4d1f-91ab-caa0beb7bff9 from datanode DatanodeRegistration(127.0.0.1:46593, datanodeUuid=d818fbea-5859-4916-8ede-7b4a208897bb, infoPort=39159, infoSecurePort=0, ipcPort=44015, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055) 2024-12-06T07:53:05,982 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc776167b283c2e89 with lease ID 0xe61d5b3456557507: from storage DS-400f7f99-b834-4d1f-91ab-caa0beb7bff9 node DatanodeRegistration(127.0.0.1:46593, datanodeUuid=d818fbea-5859-4916-8ede-7b4a208897bb, infoPort=39159, infoSecurePort=0, ipcPort=44015, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-06T07:53:05,982 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc776167b283c2e89 with lease ID 0xe61d5b3456557507: Processing first storage report for DS-fb9aaf5b-b3fd-4d87-9754-ad73d382509a from datanode DatanodeRegistration(127.0.0.1:46593, datanodeUuid=d818fbea-5859-4916-8ede-7b4a208897bb, infoPort=39159, infoSecurePort=0, ipcPort=44015, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055) 2024-12-06T07:53:05,982 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc776167b283c2e89 with lease ID 0xe61d5b3456557507: from storage DS-fb9aaf5b-b3fd-4d87-9754-ad73d382509a node DatanodeRegistration(127.0.0.1:46593, datanodeUuid=d818fbea-5859-4916-8ede-7b4a208897bb, infoPort=39159, infoSecurePort=0, ipcPort=44015, storageInfo=lv=-57;cid=testClusterID;nsid=1958764459;c=1733471584055), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T07:53:06,165 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90 2024-12-06T07:53:06,237 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-06T07:53:06,293 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=157, OpenFileDescriptor=391, MaxFileDescriptor=1048576, SystemLoadAverage=203, ProcessCount=11, AvailableMemoryMB=4152 2024-12-06T07:53:06,295 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-06T07:53:06,304 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-06T07:53:06,411 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/zookeeper_0, clientPort=61737, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-06T07:53:06,424 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=61737 2024-12-06T07:53:06,439 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:06,443 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:06,538 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:06,539 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:06,597 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:38318 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:45919:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:38318 dst: /127.0.0.1:45919 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:06,622 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-06T07:53:07,018 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:07,027 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9 with version=8 2024-12-06T07:53:07,028 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/hbase-staging 2024-12-06T07:53:07,132 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-06T07:53:07,385 INFO [Time-limited test {}] client.ConnectionUtils(128): master/aa9080fe83d2:0 server-side Connection retries=45 2024-12-06T07:53:07,396 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,396 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,401 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T07:53:07,402 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,402 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T07:53:07,539 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-06T07:53:07,599 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-06T07:53:07,608 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-06T07:53:07,612 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T07:53:07,640 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 11486 (auto-detected) 2024-12-06T07:53:07,641 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-06T07:53:07,660 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36249 2024-12-06T07:53:07,682 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:36249 connecting to ZooKeeper ensemble=127.0.0.1:61737 2024-12-06T07:53:07,721 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:362490x0, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T07:53:07,723 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:36249-0x10186c332e20000 connected 2024-12-06T07:53:07,755 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:07,759 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:07,771 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:07,776 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9, hbase.cluster.distributed=false 2024-12-06T07:53:07,804 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T07:53:07,809 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36249 2024-12-06T07:53:07,809 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36249 2024-12-06T07:53:07,809 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36249 2024-12-06T07:53:07,810 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36249 2024-12-06T07:53:07,810 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36249 2024-12-06T07:53:07,921 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/aa9080fe83d2:0 server-side Connection retries=45 2024-12-06T07:53:07,922 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,923 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,923 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T07:53:07,923 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,923 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T07:53:07,926 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T07:53:07,928 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T07:53:07,929 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:33523 2024-12-06T07:53:07,931 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:33523 connecting to ZooKeeper ensemble=127.0.0.1:61737 2024-12-06T07:53:07,932 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:07,935 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:07,942 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:335230x0, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T07:53:07,943 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:33523-0x10186c332e20001 connected 2024-12-06T07:53:07,943 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:07,949 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T07:53:07,959 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T07:53:07,962 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T07:53:07,968 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T07:53:07,969 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=33523 2024-12-06T07:53:07,970 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=33523 2024-12-06T07:53:07,971 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=33523 2024-12-06T07:53:07,973 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=33523 2024-12-06T07:53:07,976 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=33523 2024-12-06T07:53:07,994 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/aa9080fe83d2:0 server-side Connection retries=45 2024-12-06T07:53:07,994 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,994 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,995 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T07:53:07,995 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:07,995 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T07:53:07,996 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T07:53:07,996 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T07:53:07,997 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:33169 2024-12-06T07:53:07,999 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:33169 connecting to ZooKeeper ensemble=127.0.0.1:61737 2024-12-06T07:53:08,001 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:08,005 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:08,013 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:331690x0, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T07:53:08,014 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:331690x0, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:08,014 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:33169-0x10186c332e20002 connected 2024-12-06T07:53:08,014 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T07:53:08,016 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T07:53:08,017 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T07:53:08,019 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T07:53:08,022 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=33169 2024-12-06T07:53:08,024 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=33169 2024-12-06T07:53:08,025 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=33169 2024-12-06T07:53:08,027 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=33169 2024-12-06T07:53:08,030 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=33169 2024-12-06T07:53:08,046 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/aa9080fe83d2:0 server-side Connection retries=45 2024-12-06T07:53:08,046 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:08,046 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:08,046 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T07:53:08,046 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:08,047 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T07:53:08,047 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T07:53:08,047 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T07:53:08,048 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:46605 2024-12-06T07:53:08,050 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:46605 connecting to ZooKeeper ensemble=127.0.0.1:61737 2024-12-06T07:53:08,051 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:08,054 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:08,059 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:466050x0, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T07:53:08,060 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:466050x0, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:08,060 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:46605-0x10186c332e20003 connected 2024-12-06T07:53:08,061 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T07:53:08,063 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T07:53:08,064 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T07:53:08,065 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T07:53:08,066 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46605 2024-12-06T07:53:08,067 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46605 2024-12-06T07:53:08,067 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46605 2024-12-06T07:53:08,070 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46605 2024-12-06T07:53:08,071 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46605 2024-12-06T07:53:08,086 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;aa9080fe83d2:36249 2024-12-06T07:53:08,087 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:08,095 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:08,095 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:08,095 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:08,095 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:08,098 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:08,118 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T07:53:08,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T07:53:08,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T07:53:08,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,120 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-06T07:53:08,122 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/aa9080fe83d2,36249,1733471587187 from backup master directory 2024-12-06T07:53:08,124 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:08,124 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:08,125 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:08,125 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:08,125 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:08,126 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T07:53:08,126 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:08,128 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-06T07:53:08,129 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-06T07:53:08,190 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/hbase.id] with ID: e50051a1-7687-4868-940d-188951371e26 2024-12-06T07:53:08,190 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/.tmp/hbase.id 2024-12-06T07:53:08,196 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:08,196 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:08,200 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:53032 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53032 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:08,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-06T07:53:08,206 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:08,206 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/.tmp/hbase.id]:[hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/hbase.id] 2024-12-06T07:53:08,250 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:08,255 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-06T07:53:08,273 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 17ms. 2024-12-06T07:53:08,277 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,277 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,277 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,277 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,288 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:08,289 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:08,292 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:53054 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53054 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:08,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-06T07:53:08,297 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:08,312 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-06T07:53:08,314 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-06T07:53:08,319 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T07:53:08,347 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:08,347 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:08,351 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:46830 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:46593:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46830 dst: /127.0.0.1:46593 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:08,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-06T07:53:08,357 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:08,378 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store 2024-12-06T07:53:08,396 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:08,396 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:08,399 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:46858 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:46593:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46858 dst: /127.0.0.1:46593 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:08,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-06T07:53:08,404 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:08,408 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-06T07:53:08,411 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:08,412 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-06T07:53:08,412 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:08,413 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:08,414 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-06T07:53:08,414 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:08,414 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:08,415 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733471588412Disabling compacts and flushes for region at 1733471588412Disabling writes for close at 1733471588414 (+2 ms)Writing region close event to WAL at 1733471588414Closed at 1733471588414 2024-12-06T07:53:08,417 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/.initializing 2024-12-06T07:53:08,417 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/WALs/aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:08,425 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-06T07:53:08,439 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C36249%2C1733471587187, suffix=, logDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/WALs/aa9080fe83d2,36249,1733471587187, archiveDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/oldWALs, maxLogs=10 2024-12-06T07:53:08,468 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/WALs/aa9080fe83d2,36249,1733471587187/aa9080fe83d2%2C36249%2C1733471587187.1733471588443, exclude list is [], retry=0 2024-12-06T07:53:08,488 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:08,490 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2d469ba5-aa45-4caf-b546-a241f2464343,DISK] 2024-12-06T07:53:08,490 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45919,DS-45455d47-b25c-49ab-922e-4c174df95d0f,DISK] 2024-12-06T07:53:08,490 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46593,DS-400f7f99-b834-4d1f-91ab-caa0beb7bff9,DISK] 2024-12-06T07:53:08,493 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-06T07:53:08,533 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/WALs/aa9080fe83d2,36249,1733471587187/aa9080fe83d2%2C36249%2C1733471587187.1733471588443 2024-12-06T07:53:08,535 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36261:36261),(127.0.0.1/127.0.0.1:44973:44973),(127.0.0.1/127.0.0.1:39159:39159)] 2024-12-06T07:53:08,535 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-06T07:53:08,536 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:08,539 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,540 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,578 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,605 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-06T07:53:08,609 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:08,611 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:08,612 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,615 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-06T07:53:08,615 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:08,616 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T07:53:08,617 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,619 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-06T07:53:08,619 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:08,620 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T07:53:08,620 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,623 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-06T07:53:08,623 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:08,624 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T07:53:08,625 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,628 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,629 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,634 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,634 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,638 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T07:53:08,642 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:08,648 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T07:53:08,649 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=58927547, jitterRate=-0.12191112339496613}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T07:53:08,655 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733471588553Initializing all the Stores at 1733471588555 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471588556 (+1 ms)Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471588556Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471588557 (+1 ms)Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471588557Cleaning up temporary data from old regions at 1733471588635 (+78 ms)Region opened successfully at 1733471588655 (+20 ms) 2024-12-06T07:53:08,656 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-06T07:53:08,690 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6d513b59, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=aa9080fe83d2/172.17.0.2:0 2024-12-06T07:53:08,723 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-06T07:53:08,734 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-06T07:53:08,735 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-06T07:53:08,737 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-06T07:53:08,739 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-06T07:53:08,744 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 4 msec 2024-12-06T07:53:08,744 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-06T07:53:08,769 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-06T07:53:08,777 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-06T07:53:08,780 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-06T07:53:08,783 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-06T07:53:08,784 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-06T07:53:08,785 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-06T07:53:08,787 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-06T07:53:08,791 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-06T07:53:08,793 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-06T07:53:08,794 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-06T07:53:08,795 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-06T07:53:08,812 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-06T07:53:08,815 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-06T07:53:08,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:08,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:08,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:08,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:08,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,819 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,820 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,823 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=aa9080fe83d2,36249,1733471587187, sessionid=0x10186c332e20000, setting cluster-up flag (Was=false) 2024-12-06T07:53:08,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,843 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-06T07:53:08,845 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:08,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:08,857 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-06T07:53:08,859 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:08,865 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-06T07:53:08,874 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(746): ClusterId : e50051a1-7687-4868-940d-188951371e26 2024-12-06T07:53:08,874 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(746): ClusterId : e50051a1-7687-4868-940d-188951371e26 2024-12-06T07:53:08,874 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(746): ClusterId : e50051a1-7687-4868-940d-188951371e26 2024-12-06T07:53:08,877 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T07:53:08,877 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T07:53:08,877 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T07:53:08,889 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T07:53:08,889 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T07:53:08,889 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T07:53:08,889 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T07:53:08,889 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T07:53:08,889 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T07:53:08,892 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T07:53:08,892 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T07:53:08,892 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T07:53:08,893 DEBUG [RS:0;aa9080fe83d2:33523 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3e10cb64, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=aa9080fe83d2/172.17.0.2:0 2024-12-06T07:53:08,893 DEBUG [RS:1;aa9080fe83d2:33169 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7139724d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=aa9080fe83d2/172.17.0.2:0 2024-12-06T07:53:08,893 DEBUG [RS:2;aa9080fe83d2:46605 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@14cb0dda, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=aa9080fe83d2/172.17.0.2:0 2024-12-06T07:53:08,908 DEBUG [RS:0;aa9080fe83d2:33523 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;aa9080fe83d2:33523 2024-12-06T07:53:08,911 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T07:53:08,911 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T07:53:08,912 DEBUG [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T07:53:08,913 DEBUG [RS:1;aa9080fe83d2:33169 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;aa9080fe83d2:33169 2024-12-06T07:53:08,914 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;aa9080fe83d2:46605 2024-12-06T07:53:08,914 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T07:53:08,914 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T07:53:08,914 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T07:53:08,914 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T07:53:08,914 DEBUG [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T07:53:08,914 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T07:53:08,914 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,36249,1733471587187 with port=33523, startcode=1733471587881 2024-12-06T07:53:08,915 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,36249,1733471587187 with port=46605, startcode=1733471588045 2024-12-06T07:53:08,915 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,36249,1733471587187 with port=33169, startcode=1733471587993 2024-12-06T07:53:08,926 DEBUG [RS:0;aa9080fe83d2:33523 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T07:53:08,926 DEBUG [RS:2;aa9080fe83d2:46605 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T07:53:08,926 DEBUG [RS:1;aa9080fe83d2:33169 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T07:53:08,966 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-06T07:53:08,967 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48857, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T07:53:08,967 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33543, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T07:53:08,967 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49307, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T07:53:08,973 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-06T07:53:08,977 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-06T07:53:08,979 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-06T07:53:08,980 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-06T07:53:08,985 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-06T07:53:08,991 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: aa9080fe83d2,36249,1733471587187 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-06T07:53:09,000 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/aa9080fe83d2:0, corePoolSize=5, maxPoolSize=5 2024-12-06T07:53:09,000 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/aa9080fe83d2:0, corePoolSize=5, maxPoolSize=5 2024-12-06T07:53:09,000 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=5, maxPoolSize=5 2024-12-06T07:53:09,001 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=5, maxPoolSize=5 2024-12-06T07:53:09,001 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/aa9080fe83d2:0, corePoolSize=10, maxPoolSize=10 2024-12-06T07:53:09,001 DEBUG [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-06T07:53:09,001 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-06T07:53:09,001 DEBUG [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-06T07:53:09,001 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,001 WARN [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-06T07:53:09,001 WARN [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-06T07:53:09,001 WARN [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-06T07:53:09,001 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=2, maxPoolSize=2 2024-12-06T07:53:09,001 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,003 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733471619003 2024-12-06T07:53:09,005 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-06T07:53:09,006 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-06T07:53:09,007 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T07:53:09,007 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-06T07:53:09,010 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-06T07:53:09,010 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-06T07:53:09,010 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-06T07:53:09,011 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-06T07:53:09,011 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,014 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-06T07:53:09,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-06T07:53:09,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-06T07:53:09,015 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-06T07:53:09,016 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-06T07:53:09,016 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,016 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-06T07:53:09,018 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-06T07:53:09,018 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-06T07:53:09,021 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.large.0-1733471589019,5,FailOnTimeoutGroup] 2024-12-06T07:53:09,022 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.small.0-1733471589022,5,FailOnTimeoutGroup] 2024-12-06T07:53:09,022 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,022 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-06T07:53:09,023 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,024 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,024 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:09,025 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:09,031 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:53100 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53100 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:09,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-06T07:53:09,037 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:09,038 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-06T07:53:09,038 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9 2024-12-06T07:53:09,044 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:09,045 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:09,047 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:38370 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:45919:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:38370 dst: /127.0.0.1:45919 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:09,052 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-06T07:53:09,053 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:09,054 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:09,056 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-06T07:53:09,058 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-06T07:53:09,059 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,059 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:09,060 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-06T07:53:09,062 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-06T07:53:09,062 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,063 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:09,063 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-06T07:53:09,065 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-06T07:53:09,065 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,066 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:09,066 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-06T07:53:09,069 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-06T07:53:09,069 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,069 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:09,070 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-06T07:53:09,071 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740 2024-12-06T07:53:09,072 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740 2024-12-06T07:53:09,075 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-06T07:53:09,075 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-06T07:53:09,076 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T07:53:09,079 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-06T07:53:09,085 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T07:53:09,086 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67761513, jitterRate=0.009725227952003479}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T07:53:09,088 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733471589054Initializing all the Stores at 1733471589056 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471589056Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471589056Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471589056Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471589056Cleaning up temporary data from old regions at 1733471589075 (+19 ms)Region opened successfully at 1733471589088 (+13 ms) 2024-12-06T07:53:09,089 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-06T07:53:09,089 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-06T07:53:09,089 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-06T07:53:09,089 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-06T07:53:09,089 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-06T07:53:09,090 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-06T07:53:09,090 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733471589088Disabling compacts and flushes for region at 1733471589088Disabling writes for close at 1733471589089 (+1 ms)Writing region close event to WAL at 1733471589090 (+1 ms)Closed at 1733471589090 2024-12-06T07:53:09,094 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T07:53:09,094 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-06T07:53:09,101 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-06T07:53:09,102 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,36249,1733471587187 with port=33169, startcode=1733471587993 2024-12-06T07:53:09,102 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,36249,1733471587187 with port=33523, startcode=1733471587881 2024-12-06T07:53:09,102 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,36249,1733471587187 with port=46605, startcode=1733471588045 2024-12-06T07:53:09,104 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] master.ServerManager(363): Checking decommissioned status of RegionServer aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,106 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] master.ServerManager(517): Registering regionserver=aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,110 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-06T07:53:09,114 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-06T07:53:09,115 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] master.ServerManager(363): Checking decommissioned status of RegionServer aa9080fe83d2,33523,1733471587881 2024-12-06T07:53:09,115 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] master.ServerManager(517): Registering regionserver=aa9080fe83d2,33523,1733471587881 2024-12-06T07:53:09,115 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9 2024-12-06T07:53:09,116 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:39845 2024-12-06T07:53:09,116 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T07:53:09,118 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] master.ServerManager(363): Checking decommissioned status of RegionServer aa9080fe83d2,33169,1733471587993 2024-12-06T07:53:09,118 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36249 {}] master.ServerManager(517): Registering regionserver=aa9080fe83d2,33169,1733471587993 2024-12-06T07:53:09,118 DEBUG [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9 2024-12-06T07:53:09,118 DEBUG [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:39845 2024-12-06T07:53:09,118 DEBUG [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T07:53:09,120 DEBUG [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9 2024-12-06T07:53:09,121 DEBUG [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:39845 2024-12-06T07:53:09,121 DEBUG [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T07:53:09,125 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T07:53:09,126 DEBUG [RS:2;aa9080fe83d2:46605 {}] zookeeper.ZKUtil(111): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,126 DEBUG [RS:0;aa9080fe83d2:33523 {}] zookeeper.ZKUtil(111): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/aa9080fe83d2,33523,1733471587881 2024-12-06T07:53:09,127 WARN [RS:2;aa9080fe83d2:46605 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T07:53:09,127 WARN [RS:0;aa9080fe83d2:33523 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T07:53:09,127 INFO [RS:0;aa9080fe83d2:33523 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T07:53:09,127 DEBUG [RS:1;aa9080fe83d2:33169 {}] zookeeper.ZKUtil(111): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/aa9080fe83d2,33169,1733471587993 2024-12-06T07:53:09,127 INFO [RS:2;aa9080fe83d2:46605 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T07:53:09,127 WARN [RS:1;aa9080fe83d2:33169 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T07:53:09,127 INFO [RS:1;aa9080fe83d2:33169 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T07:53:09,127 DEBUG [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,33523,1733471587881 2024-12-06T07:53:09,127 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,127 DEBUG [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,33169,1733471587993 2024-12-06T07:53:09,128 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [aa9080fe83d2,33523,1733471587881] 2024-12-06T07:53:09,128 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [aa9080fe83d2,33169,1733471587993] 2024-12-06T07:53:09,129 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [aa9080fe83d2,46605,1733471588045] 2024-12-06T07:53:09,154 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T07:53:09,154 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T07:53:09,154 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T07:53:09,167 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T07:53:09,167 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T07:53:09,167 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T07:53:09,173 INFO [RS:0;aa9080fe83d2:33523 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T07:53:09,173 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,173 INFO [RS:1;aa9080fe83d2:33169 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T07:53:09,173 INFO [RS:2;aa9080fe83d2:46605 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T07:53:09,173 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,173 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,174 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T07:53:09,176 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T07:53:09,177 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T07:53:09,179 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T07:53:09,179 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T07:53:09,179 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T07:53:09,181 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,181 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,181 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,181 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,181 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,181 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,181 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,181 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,181 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/aa9080fe83d2:0, corePoolSize=2, maxPoolSize=2 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/aa9080fe83d2:0, corePoolSize=2, maxPoolSize=2 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/aa9080fe83d2:0, corePoolSize=2, maxPoolSize=2 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,182 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,183 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,183 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,183 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,183 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:09,183 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:09,183 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,183 DEBUG [RS:1;aa9080fe83d2:33169 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:09,183 DEBUG [RS:0;aa9080fe83d2:33523 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:09,183 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,183 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,183 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:09,183 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:09,184 DEBUG [RS:2;aa9080fe83d2:46605 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:09,185 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,185 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,185 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,185 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,185 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,185 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,33169,1733471587993-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T07:53:09,188 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,188 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,188 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,33523,1733471587881-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,189 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,46605,1733471588045-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T07:53:09,207 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T07:53:09,207 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T07:53:09,209 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,33169,1733471587993-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,209 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,33523,1733471587881-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,210 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,210 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,210 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.Replication(171): aa9080fe83d2,33523,1733471587881 started 2024-12-06T07:53:09,210 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.Replication(171): aa9080fe83d2,33169,1733471587993 started 2024-12-06T07:53:09,215 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T07:53:09,215 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,46605,1733471588045-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,216 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,216 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.Replication(171): aa9080fe83d2,46605,1733471588045 started 2024-12-06T07:53:09,228 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,228 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,228 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(1482): Serving as aa9080fe83d2,33523,1733471587881, RpcServer on aa9080fe83d2/172.17.0.2:33523, sessionid=0x10186c332e20001 2024-12-06T07:53:09,228 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(1482): Serving as aa9080fe83d2,33169,1733471587993, RpcServer on aa9080fe83d2/172.17.0.2:33169, sessionid=0x10186c332e20002 2024-12-06T07:53:09,229 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T07:53:09,229 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T07:53:09,229 DEBUG [RS:0;aa9080fe83d2:33523 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager aa9080fe83d2,33523,1733471587881 2024-12-06T07:53:09,229 DEBUG [RS:1;aa9080fe83d2:33169 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager aa9080fe83d2,33169,1733471587993 2024-12-06T07:53:09,229 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,33523,1733471587881' 2024-12-06T07:53:09,229 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,33169,1733471587993' 2024-12-06T07:53:09,229 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T07:53:09,229 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T07:53:09,230 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T07:53:09,230 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T07:53:09,231 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,231 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1482): Serving as aa9080fe83d2,46605,1733471588045, RpcServer on aa9080fe83d2/172.17.0.2:46605, sessionid=0x10186c332e20003 2024-12-06T07:53:09,231 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T07:53:09,231 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T07:53:09,231 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T07:53:09,231 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T07:53:09,231 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T07:53:09,231 DEBUG [RS:0;aa9080fe83d2:33523 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager aa9080fe83d2,33523,1733471587881 2024-12-06T07:53:09,231 DEBUG [RS:2;aa9080fe83d2:46605 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,231 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,33523,1733471587881' 2024-12-06T07:53:09,231 DEBUG [RS:1;aa9080fe83d2:33169 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager aa9080fe83d2,33169,1733471587993 2024-12-06T07:53:09,231 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,46605,1733471588045' 2024-12-06T07:53:09,231 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T07:53:09,231 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,33169,1733471587993' 2024-12-06T07:53:09,231 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T07:53:09,231 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T07:53:09,232 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T07:53:09,232 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T07:53:09,232 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T07:53:09,233 DEBUG [RS:1;aa9080fe83d2:33169 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T07:53:09,233 DEBUG [RS:0;aa9080fe83d2:33523 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T07:53:09,233 INFO [RS:1;aa9080fe83d2:33169 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T07:53:09,233 INFO [RS:0;aa9080fe83d2:33523 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T07:53:09,233 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T07:53:09,233 INFO [RS:1;aa9080fe83d2:33169 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T07:53:09,233 INFO [RS:0;aa9080fe83d2:33523 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T07:53:09,233 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T07:53:09,233 DEBUG [RS:2;aa9080fe83d2:46605 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,233 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,46605,1733471588045' 2024-12-06T07:53:09,233 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T07:53:09,234 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T07:53:09,234 DEBUG [RS:2;aa9080fe83d2:46605 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T07:53:09,234 INFO [RS:2;aa9080fe83d2:46605 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T07:53:09,234 INFO [RS:2;aa9080fe83d2:46605 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T07:53:09,265 WARN [aa9080fe83d2:36249 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-06T07:53:09,338 INFO [RS:2;aa9080fe83d2:46605 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-06T07:53:09,338 INFO [RS:0;aa9080fe83d2:33523 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-06T07:53:09,338 INFO [RS:1;aa9080fe83d2:33169 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-06T07:53:09,342 INFO [RS:1;aa9080fe83d2:33169 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C33169%2C1733471587993, suffix=, logDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,33169,1733471587993, archiveDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/oldWALs, maxLogs=32 2024-12-06T07:53:09,342 INFO [RS:0;aa9080fe83d2:33523 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C33523%2C1733471587881, suffix=, logDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,33523,1733471587881, archiveDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/oldWALs, maxLogs=32 2024-12-06T07:53:09,342 INFO [RS:2;aa9080fe83d2:46605 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C46605%2C1733471588045, suffix=, logDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,46605,1733471588045, archiveDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/oldWALs, maxLogs=32 2024-12-06T07:53:09,361 DEBUG [RS:2;aa9080fe83d2:46605 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,46605,1733471588045/aa9080fe83d2%2C46605%2C1733471588045.1733471589347, exclude list is [], retry=0 2024-12-06T07:53:09,361 DEBUG [RS:1;aa9080fe83d2:33169 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,33169,1733471587993/aa9080fe83d2%2C33169%2C1733471587993.1733471589347, exclude list is [], retry=0 2024-12-06T07:53:09,362 DEBUG [RS:0;aa9080fe83d2:33523 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,33523,1733471587881/aa9080fe83d2%2C33523%2C1733471587881.1733471589347, exclude list is [], retry=0 2024-12-06T07:53:09,367 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2d469ba5-aa45-4caf-b546-a241f2464343,DISK] 2024-12-06T07:53:09,367 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2d469ba5-aa45-4caf-b546-a241f2464343,DISK] 2024-12-06T07:53:09,367 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45919,DS-45455d47-b25c-49ab-922e-4c174df95d0f,DISK] 2024-12-06T07:53:09,367 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45919,DS-45455d47-b25c-49ab-922e-4c174df95d0f,DISK] 2024-12-06T07:53:09,367 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45919,DS-45455d47-b25c-49ab-922e-4c174df95d0f,DISK] 2024-12-06T07:53:09,368 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46593,DS-400f7f99-b834-4d1f-91ab-caa0beb7bff9,DISK] 2024-12-06T07:53:09,389 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2d469ba5-aa45-4caf-b546-a241f2464343,DISK] 2024-12-06T07:53:09,389 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46593,DS-400f7f99-b834-4d1f-91ab-caa0beb7bff9,DISK] 2024-12-06T07:53:09,389 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46593,DS-400f7f99-b834-4d1f-91ab-caa0beb7bff9,DISK] 2024-12-06T07:53:09,394 INFO [RS:1;aa9080fe83d2:33169 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,33169,1733471587993/aa9080fe83d2%2C33169%2C1733471587993.1733471589347 2024-12-06T07:53:09,395 INFO [RS:2;aa9080fe83d2:46605 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,46605,1733471588045/aa9080fe83d2%2C46605%2C1733471588045.1733471589347 2024-12-06T07:53:09,396 DEBUG [RS:1;aa9080fe83d2:33169 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36261:36261),(127.0.0.1/127.0.0.1:44973:44973),(127.0.0.1/127.0.0.1:39159:39159)] 2024-12-06T07:53:09,396 DEBUG [RS:2;aa9080fe83d2:46605 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:36261:36261),(127.0.0.1/127.0.0.1:39159:39159),(127.0.0.1/127.0.0.1:44973:44973)] 2024-12-06T07:53:09,396 INFO [RS:0;aa9080fe83d2:33523 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,33523,1733471587881/aa9080fe83d2%2C33523%2C1733471587881.1733471589347 2024-12-06T07:53:09,399 DEBUG [RS:0;aa9080fe83d2:33523 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:44973:44973),(127.0.0.1/127.0.0.1:36261:36261),(127.0.0.1/127.0.0.1:39159:39159)] 2024-12-06T07:53:09,518 DEBUG [aa9080fe83d2:36249 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-06T07:53:09,527 DEBUG [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(204): Hosts are {aa9080fe83d2=0} racks are {/default-rack=0} 2024-12-06T07:53:09,534 DEBUG [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-06T07:53:09,534 DEBUG [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-06T07:53:09,534 DEBUG [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-06T07:53:09,534 DEBUG [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-06T07:53:09,534 DEBUG [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-06T07:53:09,534 DEBUG [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-06T07:53:09,534 INFO [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-06T07:53:09,534 INFO [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-06T07:53:09,534 INFO [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-06T07:53:09,534 DEBUG [aa9080fe83d2:36249 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-06T07:53:09,542 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,549 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as aa9080fe83d2,46605,1733471588045, state=OPENING 2024-12-06T07:53:09,554 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-06T07:53:09,556 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:09,556 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:09,556 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:09,556 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:09,557 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:09,557 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:09,558 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:09,558 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:09,559 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-06T07:53:09,561 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=aa9080fe83d2,46605,1733471588045}] 2024-12-06T07:53:09,737 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-06T07:53:09,739 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44469, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-06T07:53:09,751 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-06T07:53:09,751 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T07:53:09,752 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-06T07:53:09,755 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C46605%2C1733471588045.meta, suffix=.meta, logDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,46605,1733471588045, archiveDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/oldWALs, maxLogs=32 2024-12-06T07:53:09,771 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,46605,1733471588045/aa9080fe83d2%2C46605%2C1733471588045.meta.1733471589757.meta, exclude list is [], retry=0 2024-12-06T07:53:09,774 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:46593,DS-400f7f99-b834-4d1f-91ab-caa0beb7bff9,DISK] 2024-12-06T07:53:09,774 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45919,DS-45455d47-b25c-49ab-922e-4c174df95d0f,DISK] 2024-12-06T07:53:09,774 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42163,DS-2d469ba5-aa45-4caf-b546-a241f2464343,DISK] 2024-12-06T07:53:09,777 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/WALs/aa9080fe83d2,46605,1733471588045/aa9080fe83d2%2C46605%2C1733471588045.meta.1733471589757.meta 2024-12-06T07:53:09,778 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:39159:39159),(127.0.0.1/127.0.0.1:36261:36261),(127.0.0.1/127.0.0.1:44973:44973)] 2024-12-06T07:53:09,778 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-06T07:53:09,779 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-06T07:53:09,782 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-06T07:53:09,787 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-06T07:53:09,816 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-06T07:53:09,817 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:09,817 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-06T07:53:09,817 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-06T07:53:09,821 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-06T07:53:09,823 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-06T07:53:09,823 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,825 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:09,825 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-06T07:53:09,827 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-06T07:53:09,827 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,828 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:09,828 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-06T07:53:09,829 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-06T07:53:09,829 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,830 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:09,830 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-06T07:53:09,832 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-06T07:53:09,832 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:09,833 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:09,833 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-06T07:53:09,834 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740 2024-12-06T07:53:09,840 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740 2024-12-06T07:53:09,842 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-06T07:53:09,842 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-06T07:53:09,843 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T07:53:09,845 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-06T07:53:09,847 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64109001, jitterRate=-0.04470144212245941}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T07:53:09,847 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-06T07:53:09,848 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733471589818Writing region info on filesystem at 1733471589818Initializing all the Stores at 1733471589820 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471589820Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471589821 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471589821Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471589821Cleaning up temporary data from old regions at 1733471589842 (+21 ms)Running coprocessor post-open hooks at 1733471589847 (+5 ms)Region opened successfully at 1733471589848 (+1 ms) 2024-12-06T07:53:09,855 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733471589729 2024-12-06T07:53:09,867 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-06T07:53:09,868 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-06T07:53:09,869 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,871 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as aa9080fe83d2,46605,1733471588045, state=OPEN 2024-12-06T07:53:09,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T07:53:09,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T07:53:09,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T07:53:09,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T07:53:09,875 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:09,875 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:09,875 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:09,875 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:09,876 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:09,881 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-06T07:53:09,881 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=aa9080fe83d2,46605,1733471588045 in 315 msec 2024-12-06T07:53:09,886 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-06T07:53:09,887 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 782 msec 2024-12-06T07:53:09,888 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T07:53:09,888 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-06T07:53:09,907 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-06T07:53:09,908 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=aa9080fe83d2,46605,1733471588045, seqNum=-1] 2024-12-06T07:53:09,928 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T07:53:09,930 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56305, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T07:53:09,953 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.0570 sec 2024-12-06T07:53:09,953 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733471589953, completionTime=-1 2024-12-06T07:53:09,955 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-06T07:53:09,956 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-06T07:53:09,979 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-06T07:53:09,979 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733471649979 2024-12-06T07:53:09,979 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733471709979 2024-12-06T07:53:09,979 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 23 msec 2024-12-06T07:53:09,981 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-06T07:53:09,987 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,36249,1733471587187-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,987 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,36249,1733471587187-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,987 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,36249,1733471587187-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,989 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-aa9080fe83d2:36249, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,989 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,989 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:09,995 DEBUG [master/aa9080fe83d2:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-06T07:53:10,015 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.889sec 2024-12-06T07:53:10,016 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-06T07:53:10,017 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-06T07:53:10,018 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-06T07:53:10,019 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-06T07:53:10,019 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-06T07:53:10,019 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,36249,1733471587187-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T07:53:10,020 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,36249,1733471587187-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-06T07:53:10,024 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-06T07:53:10,025 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-06T07:53:10,025 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,36249,1733471587187-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:10,084 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7cd2ba82, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T07:53:10,088 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-06T07:53:10,089 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-06T07:53:10,092 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request aa9080fe83d2,36249,-1 for getting cluster id 2024-12-06T07:53:10,095 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-06T07:53:10,102 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'e50051a1-7687-4868-940d-188951371e26' 2024-12-06T07:53:10,105 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-06T07:53:10,105 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "e50051a1-7687-4868-940d-188951371e26" 2024-12-06T07:53:10,105 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@47977077, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T07:53:10,105 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [aa9080fe83d2,36249,-1] 2024-12-06T07:53:10,108 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-06T07:53:10,109 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:10,111 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44806, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-06T07:53:10,114 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5d5c4d62, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T07:53:10,114 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-06T07:53:10,121 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=aa9080fe83d2,46605,1733471588045, seqNum=-1] 2024-12-06T07:53:10,122 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T07:53:10,124 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:53260, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T07:53:10,144 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:10,148 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-06T07:53:10,153 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:10,155 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@7f05b924 2024-12-06T07:53:10,156 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-06T07:53:10,158 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44822, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-06T07:53:10,164 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-06T07:53:10,174 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-06T07:53:10,177 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-06T07:53:10,179 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-06T07:53:10,180 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:10,182 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-06T07:53:10,184 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T07:53:10,191 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:10,191 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:10,197 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:38406 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:45919:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:38406 dst: /127.0.0.1:45919 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:10,202 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-06T07:53:10,203 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:10,206 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 282791b98e839ccbde3af94ee7535335, NAME => 'TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9 2024-12-06T07:53:10,212 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:10,212 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:10,215 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:53162 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:42163:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53162 dst: /127.0.0.1:42163 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:10,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-06T07:53:10,223 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:10,223 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:10,223 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 282791b98e839ccbde3af94ee7535335, disabling compactions & flushes 2024-12-06T07:53:10,224 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:10,224 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:10,224 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. after waiting 0 ms 2024-12-06T07:53:10,224 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:10,224 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:10,224 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 282791b98e839ccbde3af94ee7535335: Waiting for close lock at 1733471590223Disabling compacts and flushes for region at 1733471590223Disabling writes for close at 1733471590224 (+1 ms)Writing region close event to WAL at 1733471590224Closed at 1733471590224 2024-12-06T07:53:10,226 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-06T07:53:10,231 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733471590227"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733471590227"}]},"ts":"1733471590227"} 2024-12-06T07:53:10,237 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-06T07:53:10,239 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-06T07:53:10,242 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733471590239"}]},"ts":"1733471590239"} 2024-12-06T07:53:10,246 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-06T07:53:10,247 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {aa9080fe83d2=0} racks are {/default-rack=0} 2024-12-06T07:53:10,248 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-06T07:53:10,248 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-06T07:53:10,248 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-06T07:53:10,248 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-06T07:53:10,248 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-06T07:53:10,248 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-06T07:53:10,248 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-06T07:53:10,248 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-06T07:53:10,248 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-06T07:53:10,248 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-06T07:53:10,250 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=282791b98e839ccbde3af94ee7535335, ASSIGN}] 2024-12-06T07:53:10,252 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=282791b98e839ccbde3af94ee7535335, ASSIGN 2024-12-06T07:53:10,254 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=282791b98e839ccbde3af94ee7535335, ASSIGN; state=OFFLINE, location=aa9080fe83d2,46605,1733471588045; forceNewPlan=false, retain=false 2024-12-06T07:53:10,298 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T07:53:10,407 INFO [aa9080fe83d2:36249 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-06T07:53:10,408 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=282791b98e839ccbde3af94ee7535335, regionState=OPENING, regionLocation=aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:10,412 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=282791b98e839ccbde3af94ee7535335, ASSIGN because future has completed 2024-12-06T07:53:10,413 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 282791b98e839ccbde3af94ee7535335, server=aa9080fe83d2,46605,1733471588045}] 2024-12-06T07:53:10,507 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T07:53:10,574 INFO [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:10,574 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 282791b98e839ccbde3af94ee7535335, NAME => 'TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335.', STARTKEY => '', ENDKEY => ''} 2024-12-06T07:53:10,574 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,575 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:10,575 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,575 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,577 INFO [StoreOpener-282791b98e839ccbde3af94ee7535335-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,580 INFO [StoreOpener-282791b98e839ccbde3af94ee7535335-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 282791b98e839ccbde3af94ee7535335 columnFamilyName cf 2024-12-06T07:53:10,580 DEBUG [StoreOpener-282791b98e839ccbde3af94ee7535335-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:10,581 INFO [StoreOpener-282791b98e839ccbde3af94ee7535335-1 {}] regionserver.HStore(327): Store=282791b98e839ccbde3af94ee7535335/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T07:53:10,581 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,582 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,582 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,583 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,583 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,585 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,590 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T07:53:10,591 INFO [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 282791b98e839ccbde3af94ee7535335; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74311322, jitterRate=0.1073249876499176}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-06T07:53:10,591 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:10,592 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 282791b98e839ccbde3af94ee7535335: Running coprocessor pre-open hook at 1733471590575Writing region info on filesystem at 1733471590575Initializing all the Stores at 1733471590577 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471590577Cleaning up temporary data from old regions at 1733471590583 (+6 ms)Running coprocessor post-open hooks at 1733471590591 (+8 ms)Region opened successfully at 1733471590592 (+1 ms) 2024-12-06T07:53:10,594 INFO [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335., pid=6, masterSystemTime=1733471590567 2024-12-06T07:53:10,598 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:10,598 INFO [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:10,599 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=282791b98e839ccbde3af94ee7535335, regionState=OPEN, openSeqNum=2, regionLocation=aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:10,603 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 282791b98e839ccbde3af94ee7535335, server=aa9080fe83d2,46605,1733471588045 because future has completed 2024-12-06T07:53:10,608 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-06T07:53:10,609 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 282791b98e839ccbde3af94ee7535335, server=aa9080fe83d2,46605,1733471588045 in 192 msec 2024-12-06T07:53:10,612 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-06T07:53:10,612 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=282791b98e839ccbde3af94ee7535335, ASSIGN in 359 msec 2024-12-06T07:53:10,614 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-06T07:53:10,614 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733471590614"}]},"ts":"1733471590614"} 2024-12-06T07:53:10,617 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-06T07:53:10,618 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-06T07:53:10,621 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 450 msec 2024-12-06T07:53:10,818 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T07:53:10,819 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-06T07:53:10,819 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-06T07:53:10,820 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-06T07:53:10,826 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-06T07:53:10,827 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-06T07:53:10,827 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-06T07:53:10,835 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335., hostname=aa9080fe83d2,46605,1733471588045, seqNum=2] 2024-12-06T07:53:10,846 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-06T07:53:10,852 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-06T07:53:10,854 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T07:53:10,855 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-06T07:53:10,857 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-06T07:53:10,859 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-06T07:53:10,958 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T07:53:11,022 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=46605 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-06T07:53:11,022 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:11,026 INFO [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 282791b98e839ccbde3af94ee7535335 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-06T07:53:11,085 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335/.tmp/cf/b4016076ed5344c688e1b80aa1962158 is 36, key is row/cf:cq/1733471590838/Put/seqid=0 2024-12-06T07:53:11,091 WARN [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:11,092 WARN [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:11,095 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-2059692555_22 at /127.0.0.1:48480 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:45919:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48480 dst: /127.0.0.1:45919 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:11,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-06T07:53:11,101 WARN [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:11,101 INFO [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335/.tmp/cf/b4016076ed5344c688e1b80aa1962158 2024-12-06T07:53:11,146 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335/.tmp/cf/b4016076ed5344c688e1b80aa1962158 as hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335/cf/b4016076ed5344c688e1b80aa1962158 2024-12-06T07:53:11,157 INFO [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335/cf/b4016076ed5344c688e1b80aa1962158, entries=1, sequenceid=5, filesize=4.7 K 2024-12-06T07:53:11,164 INFO [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 282791b98e839ccbde3af94ee7535335 in 137ms, sequenceid=5, compaction requested=false 2024-12-06T07:53:11,165 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-06T07:53:11,167 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 282791b98e839ccbde3af94ee7535335: 2024-12-06T07:53:11,167 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:11,168 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T07:53:11,169 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-06T07:53:11,171 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-06T07:53:11,176 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-06T07:53:11,176 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 314 msec 2024-12-06T07:53:11,180 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 330 msec 2024-12-06T07:53:11,478 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36249 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T07:53:11,478 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-06T07:53:11,492 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-06T07:53:11,493 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-06T07:53:11,493 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:11,497 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:11,498 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:11,498 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-06T07:53:11,498 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-06T07:53:11,498 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=63992445, stopped=false 2024-12-06T07:53:11,499 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=aa9080fe83d2,36249,1733471587187 2024-12-06T07:53:11,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:11,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:11,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:11,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:11,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:11,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:11,502 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:11,502 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:11,502 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-06T07:53:11,502 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-06T07:53:11,502 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:11,502 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:11,502 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:11,503 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:11,503 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:11,503 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:11,503 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'aa9080fe83d2,33523,1733471587881' ***** 2024-12-06T07:53:11,503 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T07:53:11,503 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'aa9080fe83d2,33169,1733471587993' ***** 2024-12-06T07:53:11,503 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T07:53:11,503 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'aa9080fe83d2,46605,1733471588045' ***** 2024-12-06T07:53:11,503 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T07:53:11,504 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T07:53:11,504 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T07:53:11,504 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T07:53:11,504 INFO [RS:0;aa9080fe83d2:33523 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T07:53:11,504 INFO [RS:1;aa9080fe83d2:33169 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T07:53:11,504 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T07:53:11,504 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T07:53:11,504 INFO [RS:1;aa9080fe83d2:33169 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T07:53:11,504 INFO [RS:0;aa9080fe83d2:33523 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T07:53:11,504 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T07:53:11,504 INFO [RS:2;aa9080fe83d2:46605 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T07:53:11,504 INFO [RS:2;aa9080fe83d2:46605 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T07:53:11,504 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(959): stopping server aa9080fe83d2,33523,1733471587881 2024-12-06T07:53:11,504 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T07:53:11,505 INFO [RS:0;aa9080fe83d2:33523 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;aa9080fe83d2:33523. 2024-12-06T07:53:11,505 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(3091): Received CLOSE for 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:11,505 DEBUG [RS:0;aa9080fe83d2:33523 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:11,505 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(959): stopping server aa9080fe83d2,33169,1733471587993 2024-12-06T07:53:11,505 DEBUG [RS:0;aa9080fe83d2:33523 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:11,505 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T07:53:11,505 INFO [RS:1;aa9080fe83d2:33169 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;aa9080fe83d2:33169. 2024-12-06T07:53:11,505 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(976): stopping server aa9080fe83d2,33523,1733471587881; all regions closed. 2024-12-06T07:53:11,505 DEBUG [RS:1;aa9080fe83d2:33169 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:11,505 DEBUG [RS:1;aa9080fe83d2:33169 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:11,505 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(959): stopping server aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:11,505 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(976): stopping server aa9080fe83d2,33169,1733471587993; all regions closed. 2024-12-06T07:53:11,505 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T07:53:11,505 INFO [RS:2;aa9080fe83d2:46605 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;aa9080fe83d2:46605. 2024-12-06T07:53:11,505 DEBUG [RS:2;aa9080fe83d2:46605 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:11,505 DEBUG [RS:2;aa9080fe83d2:46605 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:11,506 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 282791b98e839ccbde3af94ee7535335, disabling compactions & flushes 2024-12-06T07:53:11,506 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T07:53:11,506 INFO [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:11,506 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T07:53:11,506 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:11,506 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T07:53:11,506 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. after waiting 0 ms 2024-12-06T07:53:11,506 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:11,506 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-06T07:53:11,506 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-06T07:53:11,507 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740, 282791b98e839ccbde3af94ee7535335=TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335.} 2024-12-06T07:53:11,506 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-06T07:53:11,507 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-06T07:53:11,507 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-06T07:53:11,507 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-06T07:53:11,507 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-06T07:53:11,507 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, 282791b98e839ccbde3af94ee7535335 2024-12-06T07:53:11,507 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-06T07:53:11,513 INFO [regionserver/aa9080fe83d2:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:11,513 INFO [regionserver/aa9080fe83d2:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:11,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_1073741828_1018 (size=93) 2024-12-06T07:53:11,514 INFO [regionserver/aa9080fe83d2:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:11,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_1073741828_1018 (size=93) 2024-12-06T07:53:11,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741828_1018 (size=93) 2024-12-06T07:53:11,517 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_1073741827_1017 (size=93) 2024-12-06T07:53:11,518 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_1073741827_1017 (size=93) 2024-12-06T07:53:11,518 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741827_1017 (size=93) 2024-12-06T07:53:11,525 DEBUG [RS:0;aa9080fe83d2:33523 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/oldWALs 2024-12-06T07:53:11,525 DEBUG [RS:1;aa9080fe83d2:33169 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/oldWALs 2024-12-06T07:53:11,525 INFO [RS:1;aa9080fe83d2:33169 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL aa9080fe83d2%2C33169%2C1733471587993:(num 1733471589347) 2024-12-06T07:53:11,525 INFO [RS:0;aa9080fe83d2:33523 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL aa9080fe83d2%2C33523%2C1733471587881:(num 1733471589347) 2024-12-06T07:53:11,525 DEBUG [RS:1;aa9080fe83d2:33169 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:11,525 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:11,525 DEBUG [RS:0;aa9080fe83d2:33523 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:11,525 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:11,526 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T07:53:11,526 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T07:53:11,526 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.ChoreService(370): Chore service for: regionserver/aa9080fe83d2:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T07:53:11,526 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.ChoreService(370): Chore service for: regionserver/aa9080fe83d2:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-06T07:53:11,526 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T07:53:11,526 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T07:53:11,526 INFO [regionserver/aa9080fe83d2:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T07:53:11,526 INFO [regionserver/aa9080fe83d2:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T07:53:11,526 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T07:53:11,526 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T07:53:11,526 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T07:53:11,526 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T07:53:11,526 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T07:53:11,526 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T07:53:11,526 INFO [RS:1;aa9080fe83d2:33169 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:33169 2024-12-06T07:53:11,527 INFO [RS:0;aa9080fe83d2:33523 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:33523 2024-12-06T07:53:11,531 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/aa9080fe83d2,33169,1733471587993 2024-12-06T07:53:11,531 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T07:53:11,531 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T07:53:11,532 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/aa9080fe83d2,33523,1733471587881 2024-12-06T07:53:11,532 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T07:53:11,533 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [aa9080fe83d2,33523,1733471587881] 2024-12-06T07:53:11,536 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/aa9080fe83d2,33523,1733471587881 already deleted, retry=false 2024-12-06T07:53:11,536 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; aa9080fe83d2,33523,1733471587881 expired; onlineServers=2 2024-12-06T07:53:11,536 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [aa9080fe83d2,33169,1733471587993] 2024-12-06T07:53:11,537 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/aa9080fe83d2,33169,1733471587993 already deleted, retry=false 2024-12-06T07:53:11,538 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; aa9080fe83d2,33169,1733471587993 expired; onlineServers=1 2024-12-06T07:53:11,541 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/default/TestHBaseWalOnEC/282791b98e839ccbde3af94ee7535335/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-06T07:53:11,544 INFO [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:11,544 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 282791b98e839ccbde3af94ee7535335: Waiting for close lock at 1733471591505Running coprocessor pre-close hooks at 1733471591505Disabling compacts and flushes for region at 1733471591505Disabling writes for close at 1733471591506 (+1 ms)Writing region close event to WAL at 1733471591508 (+2 ms)Running coprocessor post-close hooks at 1733471591542 (+34 ms)Closed at 1733471591544 (+2 ms) 2024-12-06T07:53:11,545 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335. 2024-12-06T07:53:11,551 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/info/9f36f08da8e048d29aef7d02ad94cbd5 is 153, key is TestHBaseWalOnEC,,1733471590159.282791b98e839ccbde3af94ee7535335./info:regioninfo/1733471590599/Put/seqid=0 2024-12-06T07:53:11,554 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:11,554 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:11,558 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-2059692555_22 at /127.0.0.1:45724 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:46593:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45724 dst: /127.0.0.1:46593 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:11,562 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-06T07:53:11,562 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:11,563 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/info/9f36f08da8e048d29aef7d02ad94cbd5 2024-12-06T07:53:11,591 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/ns/a2d6b722839049bab778815ad78414ef is 43, key is default/ns:d/1733471589934/Put/seqid=0 2024-12-06T07:53:11,593 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:11,593 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:11,597 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-2059692555_22 at /127.0.0.1:45744 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:46593:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45744 dst: /127.0.0.1:46593 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:11,601 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-06T07:53:11,636 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:11,636 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33169-0x10186c332e20002, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:11,636 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:11,636 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33523-0x10186c332e20001, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:11,636 INFO [RS:1;aa9080fe83d2:33169 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T07:53:11,636 INFO [RS:0;aa9080fe83d2:33523 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T07:53:11,636 INFO [RS:1;aa9080fe83d2:33169 {}] regionserver.HRegionServer(1031): Exiting; stopping=aa9080fe83d2,33169,1733471587993; zookeeper connection closed. 2024-12-06T07:53:11,636 INFO [RS:0;aa9080fe83d2:33523 {}] regionserver.HRegionServer(1031): Exiting; stopping=aa9080fe83d2,33523,1733471587881; zookeeper connection closed. 2024-12-06T07:53:11,637 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4561b23f {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4561b23f 2024-12-06T07:53:11,637 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@48f5c3ce {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@48f5c3ce 2024-12-06T07:53:11,707 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-06T07:53:11,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775741_1008 (size=1189) 2024-12-06T07:53:11,826 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775740_1008 (size=1189) 2024-12-06T07:53:11,826 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775708_1013 (size=1321) 2024-12-06T07:53:11,826 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775724_1010 (size=34) 2024-12-06T07:53:11,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775725_1010 (size=34) 2024-12-06T07:53:11,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775709_1013 (size=1321) 2024-12-06T07:53:11,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_-9223372036854775693_1015 (size=32) 2024-12-06T07:53:11,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775692_1015 (size=32) 2024-12-06T07:53:11,907 DEBUG [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-06T07:53:11,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775772_1004 (size=42) 2024-12-06T07:53:11,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775773_1004 (size=42) 2024-12-06T07:53:11,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775756_1006 (size=196) 2024-12-06T07:53:11,988 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775757_1006 (size=196) 2024-12-06T07:53:12,002 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:12,002 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/ns/a2d6b722839049bab778815ad78414ef 2024-12-06T07:53:12,030 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/table/b6aa189f14b8422db2eceb85df89b018 is 52, key is TestHBaseWalOnEC/table:state/1733471590614/Put/seqid=0 2024-12-06T07:53:12,032 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,032 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,035 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-2059692555_22 at /127.0.0.1:45834 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:46593:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45834 dst: /127.0.0.1:46593 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:12,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-06T07:53:12,039 WARN [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:12,039 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/table/b6aa189f14b8422db2eceb85df89b018 2024-12-06T07:53:12,049 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/info/9f36f08da8e048d29aef7d02ad94cbd5 as hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/info/9f36f08da8e048d29aef7d02ad94cbd5 2024-12-06T07:53:12,057 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/info/9f36f08da8e048d29aef7d02ad94cbd5, entries=10, sequenceid=11, filesize=6.5 K 2024-12-06T07:53:12,059 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/ns/a2d6b722839049bab778815ad78414ef as hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/ns/a2d6b722839049bab778815ad78414ef 2024-12-06T07:53:12,067 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/ns/a2d6b722839049bab778815ad78414ef, entries=2, sequenceid=11, filesize=5.0 K 2024-12-06T07:53:12,069 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/.tmp/table/b6aa189f14b8422db2eceb85df89b018 as hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/table/b6aa189f14b8422db2eceb85df89b018 2024-12-06T07:53:12,077 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/table/b6aa189f14b8422db2eceb85df89b018, entries=2, sequenceid=11, filesize=5.1 K 2024-12-06T07:53:12,078 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 571ms, sequenceid=11, compaction requested=false 2024-12-06T07:53:12,078 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-06T07:53:12,086 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-06T07:53:12,087 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-06T07:53:12,087 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-06T07:53:12,087 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733471591506Running coprocessor pre-close hooks at 1733471591506Disabling compacts and flushes for region at 1733471591506Disabling writes for close at 1733471591507 (+1 ms)Obtaining lock to block concurrent updates at 1733471591508 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1733471591508Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733471591509 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733471591510 (+1 ms)Flushing 1588230740/info: creating writer at 1733471591511 (+1 ms)Flushing 1588230740/info: appending metadata at 1733471591548 (+37 ms)Flushing 1588230740/info: closing flushed file at 1733471591548Flushing 1588230740/ns: creating writer at 1733471591573 (+25 ms)Flushing 1588230740/ns: appending metadata at 1733471591590 (+17 ms)Flushing 1588230740/ns: closing flushed file at 1733471591590Flushing 1588230740/table: creating writer at 1733471592011 (+421 ms)Flushing 1588230740/table: appending metadata at 1733471592028 (+17 ms)Flushing 1588230740/table: closing flushed file at 1733471592028Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@17046326: reopening flushed file at 1733471592048 (+20 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3a6106a7: reopening flushed file at 1733471592058 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6fb2968f: reopening flushed file at 1733471592067 (+9 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 571ms, sequenceid=11, compaction requested=false at 1733471592078 (+11 ms)Writing region close event to WAL at 1733471592080 (+2 ms)Running coprocessor post-close hooks at 1733471592087 (+7 ms)Closed at 1733471592087 2024-12-06T07:53:12,088 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-06T07:53:12,108 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(976): stopping server aa9080fe83d2,46605,1733471588045; all regions closed. 2024-12-06T07:53:12,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_1073741829_1019 (size=2751) 2024-12-06T07:53:12,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_1073741829_1019 (size=2751) 2024-12-06T07:53:12,112 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741829_1019 (size=2751) 2024-12-06T07:53:12,116 DEBUG [RS:2;aa9080fe83d2:46605 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/oldWALs 2024-12-06T07:53:12,116 INFO [RS:2;aa9080fe83d2:46605 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL aa9080fe83d2%2C46605%2C1733471588045.meta:.meta(num 1733471589757) 2024-12-06T07:53:12,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_1073741826_1016 (size=1298) 2024-12-06T07:53:12,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741826_1016 (size=1298) 2024-12-06T07:53:12,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_1073741826_1016 (size=1298) 2024-12-06T07:53:12,127 DEBUG [RS:2;aa9080fe83d2:46605 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/oldWALs 2024-12-06T07:53:12,127 INFO [RS:2;aa9080fe83d2:46605 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL aa9080fe83d2%2C46605%2C1733471588045:(num 1733471589347) 2024-12-06T07:53:12,127 DEBUG [RS:2;aa9080fe83d2:46605 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:12,127 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:12,127 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T07:53:12,127 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.ChoreService(370): Chore service for: regionserver/aa9080fe83d2:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T07:53:12,127 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T07:53:12,128 INFO [regionserver/aa9080fe83d2:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T07:53:12,128 INFO [RS:2;aa9080fe83d2:46605 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:46605 2024-12-06T07:53:12,130 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T07:53:12,130 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/aa9080fe83d2,46605,1733471588045 2024-12-06T07:53:12,130 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T07:53:12,131 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [aa9080fe83d2,46605,1733471588045] 2024-12-06T07:53:12,134 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/aa9080fe83d2,46605,1733471588045 already deleted, retry=false 2024-12-06T07:53:12,134 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; aa9080fe83d2,46605,1733471588045 expired; onlineServers=0 2024-12-06T07:53:12,134 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'aa9080fe83d2,36249,1733471587187' ***** 2024-12-06T07:53:12,134 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-06T07:53:12,134 INFO [M:0;aa9080fe83d2:36249 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T07:53:12,135 INFO [M:0;aa9080fe83d2:36249 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T07:53:12,135 DEBUG [M:0;aa9080fe83d2:36249 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-06T07:53:12,135 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-06T07:53:12,135 DEBUG [M:0;aa9080fe83d2:36249 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-06T07:53:12,135 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.large.0-1733471589019 {}] cleaner.HFileCleaner(306): Exit Thread[master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.large.0-1733471589019,5,FailOnTimeoutGroup] 2024-12-06T07:53:12,135 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.small.0-1733471589022 {}] cleaner.HFileCleaner(306): Exit Thread[master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.small.0-1733471589022,5,FailOnTimeoutGroup] 2024-12-06T07:53:12,135 INFO [M:0;aa9080fe83d2:36249 {}] hbase.ChoreService(370): Chore service for: master/aa9080fe83d2:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-06T07:53:12,135 INFO [M:0;aa9080fe83d2:36249 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T07:53:12,135 DEBUG [M:0;aa9080fe83d2:36249 {}] master.HMaster(1795): Stopping service threads 2024-12-06T07:53:12,135 INFO [M:0;aa9080fe83d2:36249 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-06T07:53:12,136 INFO [M:0;aa9080fe83d2:36249 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-06T07:53:12,136 INFO [M:0;aa9080fe83d2:36249 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-06T07:53:12,137 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-06T07:53:12,137 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-06T07:53:12,137 DEBUG [M:0;aa9080fe83d2:36249 {}] zookeeper.ZKUtil(347): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-06T07:53:12,137 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:12,137 WARN [M:0;aa9080fe83d2:36249 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-06T07:53:12,138 INFO [M:0;aa9080fe83d2:36249 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/.lastflushedseqids 2024-12-06T07:53:12,151 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,151 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,154 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:45860 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:46593:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45860 dst: /127.0.0.1:46593 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:12,162 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-06T07:53:12,162 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:12,163 INFO [M:0;aa9080fe83d2:36249 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-06T07:53:12,163 INFO [M:0;aa9080fe83d2:36249 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-06T07:53:12,163 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-06T07:53:12,163 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:12,163 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:12,163 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-06T07:53:12,163 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:12,163 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.84 KB heapSize=34.13 KB 2024-12-06T07:53:12,184 DEBUG [M:0;aa9080fe83d2:36249 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a69e6c3d41c34d74b94e7a694adde036 is 82, key is hbase:meta,,1/info:regioninfo/1733471589869/Put/seqid=0 2024-12-06T07:53:12,186 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,186 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,189 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:45884 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:46593:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45884 dst: /127.0.0.1:46593 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:12,193 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-06T07:53:12,193 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:12,194 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a69e6c3d41c34d74b94e7a694adde036 2024-12-06T07:53:12,218 DEBUG [M:0;aa9080fe83d2:36249 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/773cbad43e22496d85edc8b98aa86994 is 749, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733471590620/Put/seqid=0 2024-12-06T07:53:12,221 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,221 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,224 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:48554 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:45919:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48554 dst: /127.0.0.1:45919 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:12,230 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775552_1037 (size=6440) 2024-12-06T07:53:12,231 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:12,231 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.16 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/773cbad43e22496d85edc8b98aa86994 2024-12-06T07:53:12,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:12,232 INFO [RS:2;aa9080fe83d2:46605 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T07:53:12,233 INFO [RS:2;aa9080fe83d2:46605 {}] regionserver.HRegionServer(1031): Exiting; stopping=aa9080fe83d2,46605,1733471588045; zookeeper connection closed. 2024-12-06T07:53:12,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46605-0x10186c332e20003, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:12,233 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@d310b3c {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@d310b3c 2024-12-06T07:53:12,233 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-06T07:53:12,258 DEBUG [M:0;aa9080fe83d2:36249 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1e184beeab6f477b866d6eb6616f2c0a is 69, key is aa9080fe83d2,33169,1733471587993/rs:state/1733471589118/Put/seqid=0 2024-12-06T07:53:12,261 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,261 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T07:53:12,264 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1894797458_22 at /127.0.0.1:48568 [Receiving block BP-1887890605-172.17.0.2-1733471584055:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:45919:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48568 dst: /127.0.0.1:45919 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T07:53:12,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-06T07:53:12,270 WARN [M:0;aa9080fe83d2:36249 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T07:53:12,271 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1e184beeab6f477b866d6eb6616f2c0a 2024-12-06T07:53:12,281 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a69e6c3d41c34d74b94e7a694adde036 as hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/a69e6c3d41c34d74b94e7a694adde036 2024-12-06T07:53:12,289 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/a69e6c3d41c34d74b94e7a694adde036, entries=8, sequenceid=72, filesize=5.5 K 2024-12-06T07:53:12,291 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/773cbad43e22496d85edc8b98aa86994 as hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/773cbad43e22496d85edc8b98aa86994 2024-12-06T07:53:12,298 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/773cbad43e22496d85edc8b98aa86994, entries=8, sequenceid=72, filesize=6.3 K 2024-12-06T07:53:12,299 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1e184beeab6f477b866d6eb6616f2c0a as hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/1e184beeab6f477b866d6eb6616f2c0a 2024-12-06T07:53:12,306 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/1e184beeab6f477b866d6eb6616f2c0a, entries=3, sequenceid=72, filesize=5.2 K 2024-12-06T07:53:12,308 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.84 KB/27483, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 145ms, sequenceid=72, compaction requested=false 2024-12-06T07:53:12,310 INFO [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:12,310 DEBUG [M:0;aa9080fe83d2:36249 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733471592163Disabling compacts and flushes for region at 1733471592163Disabling writes for close at 1733471592163Obtaining lock to block concurrent updates at 1733471592163Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733471592163Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27483, getHeapSize=34880, getOffHeapSize=0, getCellsCount=85 at 1733471592164 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733471592165 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733471592165Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733471592183 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733471592183Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733471592201 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733471592218 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733471592218Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733471592240 (+22 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733471592258 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733471592258Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1c3bbc0b: reopening flushed file at 1733471592279 (+21 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3c615d19: reopening flushed file at 1733471592290 (+11 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@180fd209: reopening flushed file at 1733471592298 (+8 ms)Finished flush of dataSize ~26.84 KB/27483, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 145ms, sequenceid=72, compaction requested=false at 1733471592308 (+10 ms)Writing region close event to WAL at 1733471592310 (+2 ms)Closed at 1733471592310 2024-12-06T07:53:12,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46593 is added to blk_1073741825_1011 (size=32686) 2024-12-06T07:53:12,314 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42163 is added to blk_1073741825_1011 (size=32686) 2024-12-06T07:53:12,314 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45919 is added to blk_1073741825_1011 (size=32686) 2024-12-06T07:53:12,315 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T07:53:12,315 INFO [M:0;aa9080fe83d2:36249 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-06T07:53:12,315 INFO [M:0;aa9080fe83d2:36249 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36249 2024-12-06T07:53:12,315 INFO [M:0;aa9080fe83d2:36249 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T07:53:12,417 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:12,417 INFO [M:0;aa9080fe83d2:36249 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T07:53:12,417 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36249-0x10186c332e20000, quorum=127.0.0.1:61737, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:12,421 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2e59159d{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:12,423 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@a8e922f{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T07:53:12,423 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T07:53:12,424 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@24f92c39{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T07:53:12,424 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@c62369b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir/,STOPPED} 2024-12-06T07:53:12,427 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T07:53:12,427 WARN [BP-1887890605-172.17.0.2-1733471584055 heartbeating to localhost/127.0.0.1:39845 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T07:53:12,427 WARN [BP-1887890605-172.17.0.2-1733471584055 heartbeating to localhost/127.0.0.1:39845 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1887890605-172.17.0.2-1733471584055 (Datanode Uuid d818fbea-5859-4916-8ede-7b4a208897bb) service to localhost/127.0.0.1:39845 2024-12-06T07:53:12,427 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T07:53:12,429 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data5/current/BP-1887890605-172.17.0.2-1733471584055 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:12,429 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data6/current/BP-1887890605-172.17.0.2-1733471584055 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:12,430 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T07:53:12,431 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1c6b8f01{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:12,432 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11f28dd2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T07:53:12,432 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T07:53:12,432 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7fa8fa5c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T07:53:12,432 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6463ad04{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir/,STOPPED} 2024-12-06T07:53:12,433 WARN [BP-1887890605-172.17.0.2-1733471584055 heartbeating to localhost/127.0.0.1:39845 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T07:53:12,434 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T07:53:12,434 WARN [BP-1887890605-172.17.0.2-1733471584055 heartbeating to localhost/127.0.0.1:39845 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1887890605-172.17.0.2-1733471584055 (Datanode Uuid b0aa09a7-b7db-4430-af7a-e791adc4248c) service to localhost/127.0.0.1:39845 2024-12-06T07:53:12,434 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T07:53:12,434 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data3/current/BP-1887890605-172.17.0.2-1733471584055 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:12,434 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data4/current/BP-1887890605-172.17.0.2-1733471584055 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:12,435 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T07:53:12,437 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4839957b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:12,437 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5306f615{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T07:53:12,437 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T07:53:12,437 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1a2478ad{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T07:53:12,437 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@550154bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir/,STOPPED} 2024-12-06T07:53:12,439 WARN [BP-1887890605-172.17.0.2-1733471584055 heartbeating to localhost/127.0.0.1:39845 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T07:53:12,439 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T07:53:12,439 WARN [BP-1887890605-172.17.0.2-1733471584055 heartbeating to localhost/127.0.0.1:39845 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1887890605-172.17.0.2-1733471584055 (Datanode Uuid 6f74a450-2703-4638-8c9d-1cfd0de2cab4) service to localhost/127.0.0.1:39845 2024-12-06T07:53:12,439 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T07:53:12,440 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data1/current/BP-1887890605-172.17.0.2-1733471584055 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:12,440 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/cluster_e8f120fd-13e9-3cc9-ff2b-8e51ac8b9a56/data/data2/current/BP-1887890605-172.17.0.2-1733471584055 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:12,441 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T07:53:12,453 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@76e4c45c{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-06T07:53:12,454 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T07:53:12,454 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T07:53:12,454 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T07:53:12,454 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir/,STOPPED} 2024-12-06T07:53:12,464 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-06T07:53:12,497 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-06T07:53:12,506 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=84 (was 157), OpenFileDescriptor=447 (was 391) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=187 (was 203), ProcessCount=13 (was 11) - ProcessCount LEAK? -, AvailableMemoryMB=3819 (was 4152) 2024-12-06T07:53:12,514 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=84, OpenFileDescriptor=447, MaxFileDescriptor=1048576, SystemLoadAverage=187, ProcessCount=11, AvailableMemoryMB=3819 2024-12-06T07:53:12,514 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-06T07:53:12,514 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.log.dir so I do NOT create it in target/test-data/461b9302-dc9e-43b3-207c-e225cd381214 2024-12-06T07:53:12,514 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/c4417126-bbc9-b43f-6bbc-df5accd35e90/hadoop.tmp.dir so I do NOT create it in target/test-data/461b9302-dc9e-43b3-207c-e225cd381214 2024-12-06T07:53:12,514 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399, deleteOnExit=true 2024-12-06T07:53:12,514 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-06T07:53:12,515 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/test.cache.data in system properties and HBase conf 2024-12-06T07:53:12,515 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.tmp.dir in system properties and HBase conf 2024-12-06T07:53:12,515 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir in system properties and HBase conf 2024-12-06T07:53:12,515 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-06T07:53:12,515 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-06T07:53:12,515 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-06T07:53:12,515 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-06T07:53:12,515 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-06T07:53:12,515 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/nfs.dump.dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/java.io.tmpdir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-06T07:53:12,516 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-06T07:53:12,609 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:12,614 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T07:53:12,615 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T07:53:12,615 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T07:53:12,615 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T07:53:12,616 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:12,616 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@41ab5cc4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir/,AVAILABLE} 2024-12-06T07:53:12,617 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@28bcd4bd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T07:53:12,748 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4453029d{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/java.io.tmpdir/jetty-localhost-41873-hadoop-hdfs-3_4_1-tests_jar-_-any-6211028741162719792/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-06T07:53:12,749 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5ad015a3{HTTP/1.1, (http/1.1)}{localhost:41873} 2024-12-06T07:53:12,749 INFO [Time-limited test {}] server.Server(415): Started @10677ms 2024-12-06T07:53:12,851 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:12,855 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T07:53:12,857 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T07:53:12,857 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T07:53:12,857 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T07:53:12,857 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@78ab2b00{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir/,AVAILABLE} 2024-12-06T07:53:12,858 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7c7a8992{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T07:53:12,976 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@a49b909{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/java.io.tmpdir/jetty-localhost-33881-hadoop-hdfs-3_4_1-tests_jar-_-any-15133140309061298508/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:12,976 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1c513200{HTTP/1.1, (http/1.1)}{localhost:33881} 2024-12-06T07:53:12,976 INFO [Time-limited test {}] server.Server(415): Started @10904ms 2024-12-06T07:53:12,978 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T07:53:13,013 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:13,017 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T07:53:13,017 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T07:53:13,017 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T07:53:13,018 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T07:53:13,018 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@55791d09{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir/,AVAILABLE} 2024-12-06T07:53:13,018 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@505d9ca3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T07:53:13,100 WARN [Thread-530 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data1/current/BP-88630231-172.17.0.2-1733471592549/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:13,100 WARN [Thread-531 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data2/current/BP-88630231-172.17.0.2-1733471592549/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:13,118 WARN [Thread-509 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T07:53:13,122 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xea1c2228f698e74e with lease ID 0xfc7635133b46e998: Processing first storage report for DS-1a11ca70-ad94-49c2-81d5-e626d9234633 from datanode DatanodeRegistration(127.0.0.1:45937, datanodeUuid=dac1264a-8410-43a4-9005-5f20fb889398, infoPort=46163, infoSecurePort=0, ipcPort=42699, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549) 2024-12-06T07:53:13,122 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xea1c2228f698e74e with lease ID 0xfc7635133b46e998: from storage DS-1a11ca70-ad94-49c2-81d5-e626d9234633 node DatanodeRegistration(127.0.0.1:45937, datanodeUuid=dac1264a-8410-43a4-9005-5f20fb889398, infoPort=46163, infoSecurePort=0, ipcPort=42699, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-06T07:53:13,122 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xea1c2228f698e74e with lease ID 0xfc7635133b46e998: Processing first storage report for DS-d5d39f58-7c77-4bf0-887f-52b733f91a45 from datanode DatanodeRegistration(127.0.0.1:45937, datanodeUuid=dac1264a-8410-43a4-9005-5f20fb889398, infoPort=46163, infoSecurePort=0, ipcPort=42699, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549) 2024-12-06T07:53:13,122 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xea1c2228f698e74e with lease ID 0xfc7635133b46e998: from storage DS-d5d39f58-7c77-4bf0-887f-52b733f91a45 node DatanodeRegistration(127.0.0.1:45937, datanodeUuid=dac1264a-8410-43a4-9005-5f20fb889398, infoPort=46163, infoSecurePort=0, ipcPort=42699, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T07:53:13,149 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@65d24fa8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/java.io.tmpdir/jetty-localhost-43739-hadoop-hdfs-3_4_1-tests_jar-_-any-3276984254859482131/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:13,150 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@335d6851{HTTP/1.1, (http/1.1)}{localhost:43739} 2024-12-06T07:53:13,150 INFO [Time-limited test {}] server.Server(415): Started @11078ms 2024-12-06T07:53:13,151 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T07:53:13,180 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T07:53:13,183 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T07:53:13,184 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T07:53:13,184 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T07:53:13,184 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-06T07:53:13,184 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@b7fa3ef{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir/,AVAILABLE} 2024-12-06T07:53:13,185 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@660b8bbc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T07:53:13,246 WARN [Thread-565 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data3/current/BP-88630231-172.17.0.2-1733471592549/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:13,246 WARN [Thread-566 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data4/current/BP-88630231-172.17.0.2-1733471592549/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:13,263 WARN [Thread-545 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T07:53:13,266 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xca12c0b526bab3e7 with lease ID 0xfc7635133b46e999: Processing first storage report for DS-8890a6e6-655a-49e5-9330-8d0a8d511ecb from datanode DatanodeRegistration(127.0.0.1:37709, datanodeUuid=0b4abb8a-02fe-4d81-a82b-604f7a271955, infoPort=45799, infoSecurePort=0, ipcPort=41945, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549) 2024-12-06T07:53:13,266 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xca12c0b526bab3e7 with lease ID 0xfc7635133b46e999: from storage DS-8890a6e6-655a-49e5-9330-8d0a8d511ecb node DatanodeRegistration(127.0.0.1:37709, datanodeUuid=0b4abb8a-02fe-4d81-a82b-604f7a271955, infoPort=45799, infoSecurePort=0, ipcPort=41945, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T07:53:13,266 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xca12c0b526bab3e7 with lease ID 0xfc7635133b46e999: Processing first storage report for DS-a7ff105e-0b6f-49a6-ad9e-ed2c8bcd016a from datanode DatanodeRegistration(127.0.0.1:37709, datanodeUuid=0b4abb8a-02fe-4d81-a82b-604f7a271955, infoPort=45799, infoSecurePort=0, ipcPort=41945, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549) 2024-12-06T07:53:13,266 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xca12c0b526bab3e7 with lease ID 0xfc7635133b46e999: from storage DS-a7ff105e-0b6f-49a6-ad9e-ed2c8bcd016a node DatanodeRegistration(127.0.0.1:37709, datanodeUuid=0b4abb8a-02fe-4d81-a82b-604f7a271955, infoPort=45799, infoSecurePort=0, ipcPort=41945, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T07:53:13,303 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@11ec8dac{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/java.io.tmpdir/jetty-localhost-36633-hadoop-hdfs-3_4_1-tests_jar-_-any-8836624360379812080/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:13,304 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@582dea15{HTTP/1.1, (http/1.1)}{localhost:36633} 2024-12-06T07:53:13,304 INFO [Time-limited test {}] server.Server(415): Started @11232ms 2024-12-06T07:53:13,305 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T07:53:13,400 WARN [Thread-592 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data6/current/BP-88630231-172.17.0.2-1733471592549/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:13,400 WARN [Thread-591 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data5/current/BP-88630231-172.17.0.2-1733471592549/current, will proceed with Du for space computation calculation, 2024-12-06T07:53:13,418 WARN [Thread-580 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T07:53:13,421 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x18641cec4087863b with lease ID 0xfc7635133b46e99a: Processing first storage report for DS-732d898e-354d-43d7-8b26-28f86868f18f from datanode DatanodeRegistration(127.0.0.1:35709, datanodeUuid=81a92cbd-4f68-4ced-86eb-04e3045863c0, infoPort=46287, infoSecurePort=0, ipcPort=45141, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549) 2024-12-06T07:53:13,421 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x18641cec4087863b with lease ID 0xfc7635133b46e99a: from storage DS-732d898e-354d-43d7-8b26-28f86868f18f node DatanodeRegistration(127.0.0.1:35709, datanodeUuid=81a92cbd-4f68-4ced-86eb-04e3045863c0, infoPort=46287, infoSecurePort=0, ipcPort=45141, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T07:53:13,422 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x18641cec4087863b with lease ID 0xfc7635133b46e99a: Processing first storage report for DS-a793b286-6b43-4192-a920-0d53ac1f1646 from datanode DatanodeRegistration(127.0.0.1:35709, datanodeUuid=81a92cbd-4f68-4ced-86eb-04e3045863c0, infoPort=46287, infoSecurePort=0, ipcPort=45141, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549) 2024-12-06T07:53:13,422 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x18641cec4087863b with lease ID 0xfc7635133b46e99a: from storage DS-a793b286-6b43-4192-a920-0d53ac1f1646 node DatanodeRegistration(127.0.0.1:35709, datanodeUuid=81a92cbd-4f68-4ced-86eb-04e3045863c0, infoPort=46287, infoSecurePort=0, ipcPort=45141, storageInfo=lv=-57;cid=testClusterID;nsid=655518936;c=1733471592549), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T07:53:13,434 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214 2024-12-06T07:53:13,437 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/zookeeper_0, clientPort=63156, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-06T07:53:13,438 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=63156 2024-12-06T07:53:13,438 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,440 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,452 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741825_1001 (size=7) 2024-12-06T07:53:13,453 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741825_1001 (size=7) 2024-12-06T07:53:13,454 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741825_1001 (size=7) 2024-12-06T07:53:13,455 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee with version=8 2024-12-06T07:53:13,455 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:39845/user/jenkins/test-data/73ea6d44-1b67-1ef9-5470-52ce7c2762f9/hbase-staging 2024-12-06T07:53:13,457 INFO [Time-limited test {}] client.ConnectionUtils(128): master/aa9080fe83d2:0 server-side Connection retries=45 2024-12-06T07:53:13,457 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,457 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,457 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T07:53:13,457 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,457 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T07:53:13,458 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-06T07:53:13,458 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T07:53:13,459 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:38929 2024-12-06T07:53:13,460 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:38929 connecting to ZooKeeper ensemble=127.0.0.1:63156 2024-12-06T07:53:13,467 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:389290x0, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T07:53:13,468 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:38929-0x10186c34e6c0000 connected 2024-12-06T07:53:13,483 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,485 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,487 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:13,487 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee, hbase.cluster.distributed=false 2024-12-06T07:53:13,489 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T07:53:13,489 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38929 2024-12-06T07:53:13,489 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38929 2024-12-06T07:53:13,490 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38929 2024-12-06T07:53:13,490 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38929 2024-12-06T07:53:13,490 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38929 2024-12-06T07:53:13,506 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/aa9080fe83d2:0 server-side Connection retries=45 2024-12-06T07:53:13,506 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,506 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,506 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T07:53:13,506 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,506 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T07:53:13,506 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T07:53:13,506 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T07:53:13,508 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:43895 2024-12-06T07:53:13,509 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:43895 connecting to ZooKeeper ensemble=127.0.0.1:63156 2024-12-06T07:53:13,510 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,512 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:438950x0, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T07:53:13,518 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:43895-0x10186c34e6c0001 connected 2024-12-06T07:53:13,518 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:13,518 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T07:53:13,519 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T07:53:13,520 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T07:53:13,521 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T07:53:13,521 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43895 2024-12-06T07:53:13,521 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43895 2024-12-06T07:53:13,524 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43895 2024-12-06T07:53:13,525 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43895 2024-12-06T07:53:13,525 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43895 2024-12-06T07:53:13,542 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/aa9080fe83d2:0 server-side Connection retries=45 2024-12-06T07:53:13,542 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,542 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,542 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T07:53:13,542 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,542 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T07:53:13,542 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T07:53:13,542 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T07:53:13,543 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:34373 2024-12-06T07:53:13,544 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:34373 connecting to ZooKeeper ensemble=127.0.0.1:63156 2024-12-06T07:53:13,545 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,547 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,552 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:343730x0, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T07:53:13,552 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:13,552 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:34373-0x10186c34e6c0002 connected 2024-12-06T07:53:13,553 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T07:53:13,554 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T07:53:13,554 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T07:53:13,555 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T07:53:13,556 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34373 2024-12-06T07:53:13,556 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34373 2024-12-06T07:53:13,556 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34373 2024-12-06T07:53:13,557 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34373 2024-12-06T07:53:13,557 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34373 2024-12-06T07:53:13,574 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/aa9080fe83d2:0 server-side Connection retries=45 2024-12-06T07:53:13,574 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,574 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,574 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T07:53:13,574 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T07:53:13,574 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T07:53:13,574 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T07:53:13,574 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T07:53:13,575 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36189 2024-12-06T07:53:13,577 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36189 connecting to ZooKeeper ensemble=127.0.0.1:63156 2024-12-06T07:53:13,577 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,579 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,584 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:361890x0, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T07:53:13,586 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36189-0x10186c34e6c0003 connected 2024-12-06T07:53:13,586 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:13,586 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T07:53:13,587 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T07:53:13,588 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T07:53:13,592 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T07:53:13,593 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36189 2024-12-06T07:53:13,593 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36189 2024-12-06T07:53:13,594 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36189 2024-12-06T07:53:13,594 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36189 2024-12-06T07:53:13,594 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36189 2024-12-06T07:53:13,607 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;aa9080fe83d2:38929 2024-12-06T07:53:13,607 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:13,611 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:13,611 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:13,611 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:13,611 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:13,612 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:13,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T07:53:13,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T07:53:13,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,615 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,615 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T07:53:13,615 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,616 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-06T07:53:13,616 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/aa9080fe83d2,38929,1733471593457 from backup master directory 2024-12-06T07:53:13,618 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:13,618 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:13,619 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:13,619 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:13,619 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T07:53:13,619 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:13,620 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T07:53:13,627 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/hbase.id] with ID: 781a1d5a-771e-4b47-975d-da3d8ce8699e 2024-12-06T07:53:13,627 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/.tmp/hbase.id 2024-12-06T07:53:13,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741826_1002 (size=42) 2024-12-06T07:53:13,648 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741826_1002 (size=42) 2024-12-06T07:53:13,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741826_1002 (size=42) 2024-12-06T07:53:13,653 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/.tmp/hbase.id]:[hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/hbase.id] 2024-12-06T07:53:13,676 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T07:53:13,676 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-06T07:53:13,678 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-12-06T07:53:13,681 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,681 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,681 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,681 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741827_1003 (size=196) 2024-12-06T07:53:13,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741827_1003 (size=196) 2024-12-06T07:53:13,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741827_1003 (size=196) 2024-12-06T07:53:13,716 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-06T07:53:13,717 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-06T07:53:13,721 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T07:53:13,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741828_1004 (size=1189) 2024-12-06T07:53:13,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741828_1004 (size=1189) 2024-12-06T07:53:13,746 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741828_1004 (size=1189) 2024-12-06T07:53:13,748 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store 2024-12-06T07:53:13,773 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741829_1005 (size=34) 2024-12-06T07:53:13,773 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741829_1005 (size=34) 2024-12-06T07:53:13,774 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741829_1005 (size=34) 2024-12-06T07:53:13,775 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:13,775 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-06T07:53:13,775 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:13,775 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:13,775 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-06T07:53:13,776 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:13,776 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:13,776 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733471593775Disabling compacts and flushes for region at 1733471593775Disabling writes for close at 1733471593776 (+1 ms)Writing region close event to WAL at 1733471593776Closed at 1733471593776 2024-12-06T07:53:13,777 WARN [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/.initializing 2024-12-06T07:53:13,777 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/WALs/aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:13,782 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C38929%2C1733471593457, suffix=, logDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/WALs/aa9080fe83d2,38929,1733471593457, archiveDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/oldWALs, maxLogs=10 2024-12-06T07:53:13,783 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor aa9080fe83d2%2C38929%2C1733471593457.1733471593782 2024-12-06T07:53:13,808 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/WALs/aa9080fe83d2,38929,1733471593457/aa9080fe83d2%2C38929%2C1733471593457.1733471593782 2024-12-06T07:53:13,815 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45799:45799),(127.0.0.1/127.0.0.1:46163:46163),(127.0.0.1/127.0.0.1:46287:46287)] 2024-12-06T07:53:13,817 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-06T07:53:13,817 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:13,817 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,817 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,820 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,822 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-06T07:53:13,822 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,823 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:13,824 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,826 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-06T07:53:13,826 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,827 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T07:53:13,827 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,829 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-06T07:53:13,830 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,830 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T07:53:13,831 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,833 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-06T07:53:13,833 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,834 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T07:53:13,834 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,835 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,836 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,838 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,838 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,838 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T07:53:13,840 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T07:53:13,844 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T07:53:13,845 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=62580618, jitterRate=-0.06747612357139587}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T07:53:13,846 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733471593818Initializing all the Stores at 1733471593820 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471593820Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471593820Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471593820Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471593820Cleaning up temporary data from old regions at 1733471593838 (+18 ms)Region opened successfully at 1733471593846 (+8 ms) 2024-12-06T07:53:13,846 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-06T07:53:13,853 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@66488c42, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=aa9080fe83d2/172.17.0.2:0 2024-12-06T07:53:13,854 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-06T07:53:13,854 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-06T07:53:13,854 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-06T07:53:13,854 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-06T07:53:13,855 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-06T07:53:13,856 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-06T07:53:13,856 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-06T07:53:13,861 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-06T07:53:13,862 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-06T07:53:13,864 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-06T07:53:13,864 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-06T07:53:13,865 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-06T07:53:13,867 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-06T07:53:13,867 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-06T07:53:13,868 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-06T07:53:13,872 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-06T07:53:13,873 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-06T07:53:13,874 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-06T07:53:13,877 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-06T07:53:13,879 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-06T07:53:13,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:13,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:13,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:13,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:13,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,884 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=aa9080fe83d2,38929,1733471593457, sessionid=0x10186c34e6c0000, setting cluster-up flag (Was=false) 2024-12-06T07:53:13,888 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,888 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,888 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,888 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,895 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-06T07:53:13,898 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:13,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:13,909 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-06T07:53:13,910 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:13,912 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-06T07:53:13,915 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-06T07:53:13,915 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-06T07:53:13,915 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-06T07:53:13,916 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: aa9080fe83d2,38929,1733471593457 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-06T07:53:13,917 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/aa9080fe83d2:0, corePoolSize=5, maxPoolSize=5 2024-12-06T07:53:13,917 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/aa9080fe83d2:0, corePoolSize=5, maxPoolSize=5 2024-12-06T07:53:13,917 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=5, maxPoolSize=5 2024-12-06T07:53:13,917 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=5, maxPoolSize=5 2024-12-06T07:53:13,917 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/aa9080fe83d2:0, corePoolSize=10, maxPoolSize=10 2024-12-06T07:53:13,917 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:13,918 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=2, maxPoolSize=2 2024-12-06T07:53:13,918 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:13,921 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733471623921 2024-12-06T07:53:13,921 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-06T07:53:13,921 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-06T07:53:13,921 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-06T07:53:13,921 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-06T07:53:13,921 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-06T07:53:13,921 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-06T07:53:13,922 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:13,922 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-06T07:53:13,922 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-06T07:53:13,922 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-06T07:53:13,923 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-06T07:53:13,923 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-06T07:53:13,923 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T07:53:13,923 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-06T07:53:13,923 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.large.0-1733471593923,5,FailOnTimeoutGroup] 2024-12-06T07:53:13,923 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.small.0-1733471593923,5,FailOnTimeoutGroup] 2024-12-06T07:53:13,923 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:13,924 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-06T07:53:13,924 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:13,924 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:13,925 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,925 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-06T07:53:13,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741831_1007 (size=1321) 2024-12-06T07:53:13,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741831_1007 (size=1321) 2024-12-06T07:53:13,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741831_1007 (size=1321) 2024-12-06T07:53:13,940 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-06T07:53:13,940 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee 2024-12-06T07:53:13,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741832_1008 (size=32) 2024-12-06T07:53:13,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741832_1008 (size=32) 2024-12-06T07:53:13,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741832_1008 (size=32) 2024-12-06T07:53:13,951 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:13,953 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-06T07:53:13,955 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-06T07:53:13,955 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,955 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:13,956 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-06T07:53:13,957 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-06T07:53:13,957 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,958 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:13,958 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-06T07:53:13,959 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-06T07:53:13,960 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,960 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:13,960 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-06T07:53:13,962 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-06T07:53:13,962 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:13,962 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:13,963 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-06T07:53:13,963 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740 2024-12-06T07:53:13,964 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740 2024-12-06T07:53:13,966 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-06T07:53:13,966 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-06T07:53:13,967 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T07:53:13,968 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-06T07:53:13,970 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T07:53:13,971 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74869948, jitterRate=0.11564916372299194}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T07:53:13,972 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733471593951Initializing all the Stores at 1733471593952 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471593952Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471593952Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471593952Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471593952Cleaning up temporary data from old regions at 1733471593966 (+14 ms)Region opened successfully at 1733471593972 (+6 ms) 2024-12-06T07:53:13,972 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-06T07:53:13,972 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-06T07:53:13,972 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-06T07:53:13,972 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-06T07:53:13,972 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-06T07:53:13,973 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-06T07:53:13,973 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733471593972Disabling compacts and flushes for region at 1733471593972Disabling writes for close at 1733471593972Writing region close event to WAL at 1733471593973 (+1 ms)Closed at 1733471593973 2024-12-06T07:53:13,975 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T07:53:13,975 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-06T07:53:13,975 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-06T07:53:13,977 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-06T07:53:13,979 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-06T07:53:13,996 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(746): ClusterId : 781a1d5a-771e-4b47-975d-da3d8ce8699e 2024-12-06T07:53:13,996 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(746): ClusterId : 781a1d5a-771e-4b47-975d-da3d8ce8699e 2024-12-06T07:53:13,996 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(746): ClusterId : 781a1d5a-771e-4b47-975d-da3d8ce8699e 2024-12-06T07:53:13,996 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T07:53:13,996 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T07:53:13,996 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T07:53:13,999 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T07:53:14,000 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T07:53:14,000 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T07:53:14,000 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T07:53:14,000 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T07:53:14,000 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T07:53:14,003 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T07:53:14,003 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T07:53:14,003 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T07:53:14,004 DEBUG [RS:2;aa9080fe83d2:36189 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4139ee18, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=aa9080fe83d2/172.17.0.2:0 2024-12-06T07:53:14,004 DEBUG [RS:0;aa9080fe83d2:43895 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@8f2b44e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=aa9080fe83d2/172.17.0.2:0 2024-12-06T07:53:14,004 DEBUG [RS:1;aa9080fe83d2:34373 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1d22b101, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=aa9080fe83d2/172.17.0.2:0 2024-12-06T07:53:14,016 DEBUG [RS:0;aa9080fe83d2:43895 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;aa9080fe83d2:43895 2024-12-06T07:53:14,016 DEBUG [RS:1;aa9080fe83d2:34373 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;aa9080fe83d2:34373 2024-12-06T07:53:14,016 DEBUG [RS:2;aa9080fe83d2:36189 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;aa9080fe83d2:36189 2024-12-06T07:53:14,016 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T07:53:14,016 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T07:53:14,016 DEBUG [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T07:53:14,016 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T07:53:14,016 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T07:53:14,016 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T07:53:14,016 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T07:53:14,016 DEBUG [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T07:53:14,016 DEBUG [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T07:53:14,017 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,38929,1733471593457 with port=43895, startcode=1733471593506 2024-12-06T07:53:14,017 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,38929,1733471593457 with port=36189, startcode=1733471593573 2024-12-06T07:53:14,017 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(2659): reportForDuty to master=aa9080fe83d2,38929,1733471593457 with port=34373, startcode=1733471593541 2024-12-06T07:53:14,018 DEBUG [RS:1;aa9080fe83d2:34373 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T07:53:14,018 DEBUG [RS:0;aa9080fe83d2:43895 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T07:53:14,018 DEBUG [RS:2;aa9080fe83d2:36189 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T07:53:14,020 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43951, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T07:53:14,020 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:53541, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T07:53:14,020 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:40615, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T07:53:14,021 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38929 {}] master.ServerManager(363): Checking decommissioned status of RegionServer aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:14,021 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38929 {}] master.ServerManager(517): Registering regionserver=aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:14,023 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38929 {}] master.ServerManager(363): Checking decommissioned status of RegionServer aa9080fe83d2,34373,1733471593541 2024-12-06T07:53:14,024 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38929 {}] master.ServerManager(517): Registering regionserver=aa9080fe83d2,34373,1733471593541 2024-12-06T07:53:14,024 DEBUG [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee 2024-12-06T07:53:14,024 DEBUG [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:32919 2024-12-06T07:53:14,024 DEBUG [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T07:53:14,026 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38929 {}] master.ServerManager(363): Checking decommissioned status of RegionServer aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,026 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38929 {}] master.ServerManager(517): Registering regionserver=aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,026 DEBUG [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee 2024-12-06T07:53:14,026 DEBUG [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:32919 2024-12-06T07:53:14,026 DEBUG [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T07:53:14,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T07:53:14,027 DEBUG [RS:0;aa9080fe83d2:43895 {}] zookeeper.ZKUtil(111): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:14,027 WARN [RS:0;aa9080fe83d2:43895 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T07:53:14,027 INFO [RS:0;aa9080fe83d2:43895 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T07:53:14,027 DEBUG [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:14,028 DEBUG [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee 2024-12-06T07:53:14,028 DEBUG [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:32919 2024-12-06T07:53:14,028 DEBUG [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T07:53:14,034 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [aa9080fe83d2,34373,1733471593541] 2024-12-06T07:53:14,035 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [aa9080fe83d2,43895,1733471593506] 2024-12-06T07:53:14,035 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T07:53:14,036 DEBUG [RS:1;aa9080fe83d2:34373 {}] zookeeper.ZKUtil(111): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/aa9080fe83d2,34373,1733471593541 2024-12-06T07:53:14,036 WARN [RS:1;aa9080fe83d2:34373 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T07:53:14,036 INFO [RS:1;aa9080fe83d2:34373 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T07:53:14,036 DEBUG [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,34373,1733471593541 2024-12-06T07:53:14,036 DEBUG [RS:2;aa9080fe83d2:36189 {}] zookeeper.ZKUtil(111): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,036 WARN [RS:2;aa9080fe83d2:36189 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T07:53:14,036 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [aa9080fe83d2,36189,1733471593573] 2024-12-06T07:53:14,036 INFO [RS:2;aa9080fe83d2:36189 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T07:53:14,036 DEBUG [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,037 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T07:53:14,040 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T07:53:14,041 INFO [RS:0;aa9080fe83d2:43895 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T07:53:14,041 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,041 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T07:53:14,043 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T07:53:14,044 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T07:53:14,044 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,044 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,044 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,044 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,045 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,045 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,045 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/aa9080fe83d2:0, corePoolSize=2, maxPoolSize=2 2024-12-06T07:53:14,045 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T07:53:14,045 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,045 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,045 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,045 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,046 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,046 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,046 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:14,046 DEBUG [RS:0;aa9080fe83d2:43895 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:14,048 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T07:53:14,048 INFO [RS:1;aa9080fe83d2:34373 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T07:53:14,049 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,049 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,049 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,049 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,049 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,049 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,049 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,43895,1733471593506-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T07:53:14,050 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T07:53:14,051 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T07:53:14,051 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T07:53:14,052 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,052 INFO [RS:2;aa9080fe83d2:36189 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,052 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,052 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/aa9080fe83d2:0, corePoolSize=2, maxPoolSize=2 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,052 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,053 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,053 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,053 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,053 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:14,053 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T07:53:14,053 DEBUG [RS:1;aa9080fe83d2:34373 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:14,053 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,053 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,053 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,053 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,053 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/aa9080fe83d2:0, corePoolSize=2, maxPoolSize=2 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/aa9080fe83d2:0, corePoolSize=1, maxPoolSize=1 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:14,054 DEBUG [RS:2;aa9080fe83d2:36189 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0, corePoolSize=3, maxPoolSize=3 2024-12-06T07:53:14,057 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,34373,1733471593541-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T07:53:14,057 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,36189,1733471593573-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T07:53:14,071 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T07:53:14,072 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,43895,1733471593506-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,072 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,072 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.Replication(171): aa9080fe83d2,43895,1733471593506 started 2024-12-06T07:53:14,073 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T07:53:14,073 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,36189,1733471593573-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,074 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,074 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.Replication(171): aa9080fe83d2,36189,1733471593573 started 2024-12-06T07:53:14,079 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T07:53:14,080 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,34373,1733471593541-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,080 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,080 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.Replication(171): aa9080fe83d2,34373,1733471593541 started 2024-12-06T07:53:14,089 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,089 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1482): Serving as aa9080fe83d2,36189,1733471593573, RpcServer on aa9080fe83d2/172.17.0.2:36189, sessionid=0x10186c34e6c0003 2024-12-06T07:53:14,089 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T07:53:14,089 DEBUG [RS:2;aa9080fe83d2:36189 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,089 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,36189,1733471593573' 2024-12-06T07:53:14,089 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T07:53:14,090 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T07:53:14,091 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T07:53:14,091 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T07:53:14,091 DEBUG [RS:2;aa9080fe83d2:36189 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,091 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,36189,1733471593573' 2024-12-06T07:53:14,091 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T07:53:14,092 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T07:53:14,093 DEBUG [RS:2;aa9080fe83d2:36189 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T07:53:14,093 INFO [RS:2;aa9080fe83d2:36189 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T07:53:14,093 INFO [RS:2;aa9080fe83d2:36189 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T07:53:14,094 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,094 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1482): Serving as aa9080fe83d2,43895,1733471593506, RpcServer on aa9080fe83d2/172.17.0.2:43895, sessionid=0x10186c34e6c0001 2024-12-06T07:53:14,094 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T07:53:14,094 DEBUG [RS:0;aa9080fe83d2:43895 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:14,094 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,43895,1733471593506' 2024-12-06T07:53:14,094 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T07:53:14,095 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T07:53:14,096 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T07:53:14,096 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T07:53:14,096 DEBUG [RS:0;aa9080fe83d2:43895 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:14,096 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,43895,1733471593506' 2024-12-06T07:53:14,096 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T07:53:14,097 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T07:53:14,097 DEBUG [RS:0;aa9080fe83d2:43895 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T07:53:14,097 INFO [RS:0;aa9080fe83d2:43895 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T07:53:14,097 INFO [RS:0;aa9080fe83d2:43895 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T07:53:14,100 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,100 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(1482): Serving as aa9080fe83d2,34373,1733471593541, RpcServer on aa9080fe83d2/172.17.0.2:34373, sessionid=0x10186c34e6c0002 2024-12-06T07:53:14,100 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T07:53:14,100 DEBUG [RS:1;aa9080fe83d2:34373 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager aa9080fe83d2,34373,1733471593541 2024-12-06T07:53:14,100 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,34373,1733471593541' 2024-12-06T07:53:14,100 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T07:53:14,101 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T07:53:14,101 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T07:53:14,101 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T07:53:14,101 DEBUG [RS:1;aa9080fe83d2:34373 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager aa9080fe83d2,34373,1733471593541 2024-12-06T07:53:14,102 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'aa9080fe83d2,34373,1733471593541' 2024-12-06T07:53:14,102 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T07:53:14,102 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T07:53:14,102 DEBUG [RS:1;aa9080fe83d2:34373 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T07:53:14,102 INFO [RS:1;aa9080fe83d2:34373 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T07:53:14,102 INFO [RS:1;aa9080fe83d2:34373 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T07:53:14,129 WARN [aa9080fe83d2:38929 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-06T07:53:14,195 INFO [RS:2;aa9080fe83d2:36189 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C36189%2C1733471593573, suffix=, logDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,36189,1733471593573, archiveDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/oldWALs, maxLogs=32 2024-12-06T07:53:14,197 INFO [RS:2;aa9080fe83d2:36189 {}] monitor.StreamSlowMonitor(122): New stream slow monitor aa9080fe83d2%2C36189%2C1733471593573.1733471594197 2024-12-06T07:53:14,200 INFO [RS:0;aa9080fe83d2:43895 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C43895%2C1733471593506, suffix=, logDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,43895,1733471593506, archiveDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/oldWALs, maxLogs=32 2024-12-06T07:53:14,202 INFO [RS:0;aa9080fe83d2:43895 {}] monitor.StreamSlowMonitor(122): New stream slow monitor aa9080fe83d2%2C43895%2C1733471593506.1733471594202 2024-12-06T07:53:14,205 INFO [RS:1;aa9080fe83d2:34373 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C34373%2C1733471593541, suffix=, logDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,34373,1733471593541, archiveDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/oldWALs, maxLogs=32 2024-12-06T07:53:14,207 INFO [RS:1;aa9080fe83d2:34373 {}] monitor.StreamSlowMonitor(122): New stream slow monitor aa9080fe83d2%2C34373%2C1733471593541.1733471594207 2024-12-06T07:53:14,209 INFO [RS:2;aa9080fe83d2:36189 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,36189,1733471593573/aa9080fe83d2%2C36189%2C1733471593573.1733471594197 2024-12-06T07:53:14,215 DEBUG [RS:2;aa9080fe83d2:36189 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46287:46287),(127.0.0.1/127.0.0.1:46163:46163),(127.0.0.1/127.0.0.1:45799:45799)] 2024-12-06T07:53:14,215 INFO [RS:0;aa9080fe83d2:43895 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,43895,1733471593506/aa9080fe83d2%2C43895%2C1733471593506.1733471594202 2024-12-06T07:53:14,223 INFO [RS:1;aa9080fe83d2:34373 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,34373,1733471593541/aa9080fe83d2%2C34373%2C1733471593541.1733471594207 2024-12-06T07:53:14,225 DEBUG [RS:0;aa9080fe83d2:43895 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46163:46163),(127.0.0.1/127.0.0.1:45799:45799),(127.0.0.1/127.0.0.1:46287:46287)] 2024-12-06T07:53:14,229 DEBUG [RS:1;aa9080fe83d2:34373 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45799:45799),(127.0.0.1/127.0.0.1:46163:46163),(127.0.0.1/127.0.0.1:46287:46287)] 2024-12-06T07:53:14,379 DEBUG [aa9080fe83d2:38929 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-06T07:53:14,379 DEBUG [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(204): Hosts are {aa9080fe83d2=0} racks are {/default-rack=0} 2024-12-06T07:53:14,382 DEBUG [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-06T07:53:14,382 DEBUG [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-06T07:53:14,382 DEBUG [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-06T07:53:14,382 DEBUG [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-06T07:53:14,382 DEBUG [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-06T07:53:14,382 DEBUG [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-06T07:53:14,382 INFO [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-06T07:53:14,382 INFO [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-06T07:53:14,382 INFO [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-06T07:53:14,382 DEBUG [aa9080fe83d2:38929 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-06T07:53:14,383 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,385 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as aa9080fe83d2,36189,1733471593573, state=OPENING 2024-12-06T07:53:14,387 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-06T07:53:14,388 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:14,388 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:14,388 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:14,388 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:14,389 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-06T07:53:14,389 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:14,389 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=aa9080fe83d2,36189,1733471593573}] 2024-12-06T07:53:14,389 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:14,389 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:14,389 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:14,543 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-06T07:53:14,545 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48743, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-06T07:53:14,551 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-06T07:53:14,551 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T07:53:14,554 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=aa9080fe83d2%2C36189%2C1733471593573.meta, suffix=.meta, logDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,36189,1733471593573, archiveDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/oldWALs, maxLogs=32 2024-12-06T07:53:14,556 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor aa9080fe83d2%2C36189%2C1733471593573.meta.1733471594555.meta 2024-12-06T07:53:14,570 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/WALs/aa9080fe83d2,36189,1733471593573/aa9080fe83d2%2C36189%2C1733471593573.meta.1733471594555.meta 2024-12-06T07:53:14,574 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46163:46163),(127.0.0.1/127.0.0.1:45799:45799),(127.0.0.1/127.0.0.1:46287:46287)] 2024-12-06T07:53:14,584 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-06T07:53:14,585 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-06T07:53:14,585 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-06T07:53:14,585 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-06T07:53:14,585 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-06T07:53:14,585 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:14,585 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-06T07:53:14,585 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-06T07:53:14,588 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-06T07:53:14,589 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-06T07:53:14,589 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:14,590 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:14,590 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-06T07:53:14,591 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-06T07:53:14,591 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:14,592 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:14,592 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-06T07:53:14,593 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-06T07:53:14,593 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:14,594 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:14,594 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-06T07:53:14,595 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-06T07:53:14,595 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:14,596 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T07:53:14,596 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-06T07:53:14,597 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740 2024-12-06T07:53:14,599 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740 2024-12-06T07:53:14,601 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-06T07:53:14,601 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-06T07:53:14,601 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T07:53:14,603 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-06T07:53:14,605 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63580515, jitterRate=-0.052576497197151184}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T07:53:14,605 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-06T07:53:14,606 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733471594586Writing region info on filesystem at 1733471594586Initializing all the Stores at 1733471594587 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471594587Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471594588 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471594588Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733471594588Cleaning up temporary data from old regions at 1733471594601 (+13 ms)Running coprocessor post-open hooks at 1733471594605 (+4 ms)Region opened successfully at 1733471594606 (+1 ms) 2024-12-06T07:53:14,608 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733471594543 2024-12-06T07:53:14,612 DEBUG [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-06T07:53:14,612 INFO [RS_OPEN_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-06T07:53:14,613 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,614 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as aa9080fe83d2,36189,1733471593573, state=OPEN 2024-12-06T07:53:14,616 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T07:53:14,616 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T07:53:14,616 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T07:53:14,616 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T07:53:14,616 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:14,616 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:14,616 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:14,616 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:14,616 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T07:53:14,620 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-06T07:53:14,620 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=aa9080fe83d2,36189,1733471593573 in 227 msec 2024-12-06T07:53:14,624 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-06T07:53:14,624 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 645 msec 2024-12-06T07:53:14,625 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T07:53:14,625 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-06T07:53:14,627 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-06T07:53:14,627 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=aa9080fe83d2,36189,1733471593573, seqNum=-1] 2024-12-06T07:53:14,627 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T07:53:14,629 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46691, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T07:53:14,636 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 722 msec 2024-12-06T07:53:14,637 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733471594637, completionTime=-1 2024-12-06T07:53:14,637 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-06T07:53:14,637 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-06T07:53:14,639 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-06T07:53:14,639 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733471654639 2024-12-06T07:53:14,639 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733471714639 2024-12-06T07:53:14,639 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-12-06T07:53:14,640 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-06T07:53:14,640 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,38929,1733471593457-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,640 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,38929,1733471593457-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,640 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,38929,1733471593457-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,640 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-aa9080fe83d2:38929, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,641 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,641 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,643 DEBUG [master/aa9080fe83d2:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-06T07:53:14,646 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.027sec 2024-12-06T07:53:14,647 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-06T07:53:14,647 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-06T07:53:14,647 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-06T07:53:14,647 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-06T07:53:14,647 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-06T07:53:14,647 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,38929,1733471593457-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T07:53:14,647 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,38929,1733471593457-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-06T07:53:14,649 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-06T07:53:14,649 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-06T07:53:14,649 INFO [master/aa9080fe83d2:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=aa9080fe83d2,38929,1733471593457-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T07:53:14,697 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4d5d2159, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T07:53:14,697 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request aa9080fe83d2,38929,-1 for getting cluster id 2024-12-06T07:53:14,697 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-06T07:53:14,699 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '781a1d5a-771e-4b47-975d-da3d8ce8699e' 2024-12-06T07:53:14,699 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-06T07:53:14,699 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "781a1d5a-771e-4b47-975d-da3d8ce8699e" 2024-12-06T07:53:14,700 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5af1e308, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T07:53:14,700 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [aa9080fe83d2,38929,-1] 2024-12-06T07:53:14,700 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-06T07:53:14,700 DEBUG [RPCClient-NioEventLoopGroup-6-5 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:14,702 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58282, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-06T07:53:14,703 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1f519320, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T07:53:14,703 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-06T07:53:14,705 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=aa9080fe83d2,36189,1733471593573, seqNum=-1] 2024-12-06T07:53:14,705 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T07:53:14,706 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55542, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T07:53:14,709 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:14,709 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-06T07:53:14,710 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.AsyncConnectionImpl(321): The fetched master address is aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:14,710 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@245e24c5 2024-12-06T07:53:14,711 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-06T07:53:14,712 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58284, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-06T07:53:14,713 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-06T07:53:14,716 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-06T07:53:14,718 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-06T07:53:14,718 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:14,718 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-06T07:53:14,719 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-06T07:53:14,719 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T07:53:14,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741837_1013 (size=392) 2024-12-06T07:53:14,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741837_1013 (size=392) 2024-12-06T07:53:14,730 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741837_1013 (size=392) 2024-12-06T07:53:14,732 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 9bf2b886a5e12312fc292a4e0e0b2bc9, NAME => 'TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee 2024-12-06T07:53:14,741 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741838_1014 (size=51) 2024-12-06T07:53:14,741 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741838_1014 (size=51) 2024-12-06T07:53:14,741 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741838_1014 (size=51) 2024-12-06T07:53:14,742 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:14,742 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 9bf2b886a5e12312fc292a4e0e0b2bc9, disabling compactions & flushes 2024-12-06T07:53:14,742 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:14,742 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:14,742 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. after waiting 0 ms 2024-12-06T07:53:14,742 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:14,742 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:14,743 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 9bf2b886a5e12312fc292a4e0e0b2bc9: Waiting for close lock at 1733471594742Disabling compacts and flushes for region at 1733471594742Disabling writes for close at 1733471594742Writing region close event to WAL at 1733471594742Closed at 1733471594742 2024-12-06T07:53:14,745 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-06T07:53:14,745 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733471594745"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733471594745"}]},"ts":"1733471594745"} 2024-12-06T07:53:14,748 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-06T07:53:14,749 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-06T07:53:14,750 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733471594750"}]},"ts":"1733471594750"} 2024-12-06T07:53:14,752 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-06T07:53:14,753 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {aa9080fe83d2=0} racks are {/default-rack=0} 2024-12-06T07:53:14,753 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-06T07:53:14,753 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-06T07:53:14,753 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-06T07:53:14,753 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-06T07:53:14,753 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-06T07:53:14,753 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-06T07:53:14,753 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-06T07:53:14,753 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-06T07:53:14,753 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-06T07:53:14,753 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-06T07:53:14,754 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=9bf2b886a5e12312fc292a4e0e0b2bc9, ASSIGN}] 2024-12-06T07:53:14,756 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=9bf2b886a5e12312fc292a4e0e0b2bc9, ASSIGN 2024-12-06T07:53:14,758 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=9bf2b886a5e12312fc292a4e0e0b2bc9, ASSIGN; state=OFFLINE, location=aa9080fe83d2,43895,1733471593506; forceNewPlan=false, retain=false 2024-12-06T07:53:14,828 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T07:53:14,908 INFO [aa9080fe83d2:38929 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-06T07:53:14,909 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=9bf2b886a5e12312fc292a4e0e0b2bc9, regionState=OPENING, regionLocation=aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:14,913 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=9bf2b886a5e12312fc292a4e0e0b2bc9, ASSIGN because future has completed 2024-12-06T07:53:14,913 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 9bf2b886a5e12312fc292a4e0e0b2bc9, server=aa9080fe83d2,43895,1733471593506}] 2024-12-06T07:53:15,038 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T07:53:15,068 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-06T07:53:15,070 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35087, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-06T07:53:15,074 INFO [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,075 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 9bf2b886a5e12312fc292a4e0e0b2bc9, NAME => 'TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9.', STARTKEY => '', ENDKEY => ''} 2024-12-06T07:53:15,075 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,075 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T07:53:15,075 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,075 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,077 INFO [StoreOpener-9bf2b886a5e12312fc292a4e0e0b2bc9-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,078 INFO [StoreOpener-9bf2b886a5e12312fc292a4e0e0b2bc9-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 9bf2b886a5e12312fc292a4e0e0b2bc9 columnFamilyName cf 2024-12-06T07:53:15,079 DEBUG [StoreOpener-9bf2b886a5e12312fc292a4e0e0b2bc9-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T07:53:15,079 INFO [StoreOpener-9bf2b886a5e12312fc292a4e0e0b2bc9-1 {}] regionserver.HStore(327): Store=9bf2b886a5e12312fc292a4e0e0b2bc9/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T07:53:15,079 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,080 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,081 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,081 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,081 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,083 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,085 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T07:53:15,086 INFO [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 9bf2b886a5e12312fc292a4e0e0b2bc9; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=59761174, jitterRate=-0.10948911309242249}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-06T07:53:15,086 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,087 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 9bf2b886a5e12312fc292a4e0e0b2bc9: Running coprocessor pre-open hook at 1733471595075Writing region info on filesystem at 1733471595075Initializing all the Stores at 1733471595076 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733471595077 (+1 ms)Cleaning up temporary data from old regions at 1733471595081 (+4 ms)Running coprocessor post-open hooks at 1733471595086 (+5 ms)Region opened successfully at 1733471595087 (+1 ms) 2024-12-06T07:53:15,088 INFO [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9., pid=6, masterSystemTime=1733471595067 2024-12-06T07:53:15,091 DEBUG [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,091 INFO [RS_OPEN_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,092 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=9bf2b886a5e12312fc292a4e0e0b2bc9, regionState=OPEN, openSeqNum=2, regionLocation=aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:15,096 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 9bf2b886a5e12312fc292a4e0e0b2bc9, server=aa9080fe83d2,43895,1733471593506 because future has completed 2024-12-06T07:53:15,101 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-06T07:53:15,101 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 9bf2b886a5e12312fc292a4e0e0b2bc9, server=aa9080fe83d2,43895,1733471593506 in 184 msec 2024-12-06T07:53:15,105 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-06T07:53:15,105 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=9bf2b886a5e12312fc292a4e0e0b2bc9, ASSIGN in 347 msec 2024-12-06T07:53:15,106 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-06T07:53:15,106 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733471595106"}]},"ts":"1733471595106"} 2024-12-06T07:53:15,109 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-06T07:53:15,110 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-06T07:53:15,113 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 397 msec 2024-12-06T07:53:15,320 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-06T07:53:15,328 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-06T07:53:15,348 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T07:53:15,348 INFO [RPCClient-NioEventLoopGroup-6-8 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-06T07:53:15,348 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-06T07:53:15,349 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-06T07:53:15,353 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-06T07:53:15,353 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-06T07:53:15,353 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-06T07:53:15,357 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9., hostname=aa9080fe83d2,43895,1733471593506, seqNum=2] 2024-12-06T07:53:15,357 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T07:53:15,360 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49712, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T07:53:15,363 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-06T07:53:15,365 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-06T07:53:15,366 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T07:53:15,367 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-06T07:53:15,368 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-06T07:53:15,369 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-06T07:53:15,380 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-06T07:53:15,381 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-06T07:53:15,478 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T07:53:15,523 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=43895 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-06T07:53:15,524 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,524 INFO [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 9bf2b886a5e12312fc292a4e0e0b2bc9 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-06T07:53:15,542 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9/.tmp/cf/6e80ec1ecc6049349cef9846e1ac9779 is 36, key is row/cf:cq/1733471595360/Put/seqid=0 2024-12-06T07:53:15,550 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741839_1015 (size=4787) 2024-12-06T07:53:15,550 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741839_1015 (size=4787) 2024-12-06T07:53:15,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741839_1015 (size=4787) 2024-12-06T07:53:15,552 INFO [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9/.tmp/cf/6e80ec1ecc6049349cef9846e1ac9779 2024-12-06T07:53:15,561 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9/.tmp/cf/6e80ec1ecc6049349cef9846e1ac9779 as hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9/cf/6e80ec1ecc6049349cef9846e1ac9779 2024-12-06T07:53:15,570 INFO [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9/cf/6e80ec1ecc6049349cef9846e1ac9779, entries=1, sequenceid=5, filesize=4.7 K 2024-12-06T07:53:15,571 INFO [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 9bf2b886a5e12312fc292a4e0e0b2bc9 in 47ms, sequenceid=5, compaction requested=false 2024-12-06T07:53:15,571 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 9bf2b886a5e12312fc292a4e0e0b2bc9: 2024-12-06T07:53:15,572 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,572 DEBUG [RS_FLUSH_OPERATIONS-regionserver/aa9080fe83d2:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-06T07:53:15,572 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-06T07:53:15,577 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-06T07:53:15,578 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 205 msec 2024-12-06T07:53:15,581 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 216 msec 2024-12-06T07:53:15,687 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38929 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T07:53:15,688 INFO [RPCClient-NioEventLoopGroup-6-8 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-06T07:53:15,692 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-06T07:53:15,692 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-06T07:53:15,693 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:15,693 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,693 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,693 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-06T07:53:15,693 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-06T07:53:15,693 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=443420440, stopped=false 2024-12-06T07:53:15,693 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=aa9080fe83d2,38929,1733471593457 2024-12-06T07:53:15,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:15,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:15,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:15,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:15,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:15,695 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-06T07:53:15,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:15,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T07:53:15,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:15,695 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-06T07:53:15,696 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:15,696 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,696 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:15,696 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:15,696 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:15,696 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'aa9080fe83d2,43895,1733471593506' ***** 2024-12-06T07:53:15,696 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T07:53:15,696 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T07:53:15,696 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'aa9080fe83d2,34373,1733471593541' ***** 2024-12-06T07:53:15,696 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T07:53:15,696 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'aa9080fe83d2,36189,1733471593573' ***** 2024-12-06T07:53:15,696 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T07:53:15,697 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T07:53:15,697 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T07:53:15,697 INFO [RS:0;aa9080fe83d2:43895 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T07:53:15,697 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T07:53:15,697 INFO [RS:0;aa9080fe83d2:43895 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T07:53:15,697 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(3091): Received CLOSE for 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,697 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T07:53:15,697 INFO [RS:2;aa9080fe83d2:36189 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T07:53:15,697 INFO [RS:1;aa9080fe83d2:34373 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T07:53:15,697 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T07:53:15,697 INFO [RS:1;aa9080fe83d2:34373 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T07:53:15,697 INFO [RS:2;aa9080fe83d2:36189 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T07:53:15,697 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(959): stopping server aa9080fe83d2,34373,1733471593541 2024-12-06T07:53:15,697 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(959): stopping server aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:15,697 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T07:53:15,697 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T07:53:15,697 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T07:53:15,697 INFO [RS:1;aa9080fe83d2:34373 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;aa9080fe83d2:34373. 2024-12-06T07:53:15,697 INFO [RS:2;aa9080fe83d2:36189 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;aa9080fe83d2:36189. 2024-12-06T07:53:15,698 DEBUG [RS:1;aa9080fe83d2:34373 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:15,698 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(959): stopping server aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:15,698 DEBUG [RS:2;aa9080fe83d2:36189 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:15,698 DEBUG [RS:1;aa9080fe83d2:34373 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,698 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T07:53:15,698 DEBUG [RS:2;aa9080fe83d2:36189 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,698 INFO [RS:0;aa9080fe83d2:43895 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;aa9080fe83d2:43895. 2024-12-06T07:53:15,698 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(976): stopping server aa9080fe83d2,34373,1733471593541; all regions closed. 2024-12-06T07:53:15,698 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 9bf2b886a5e12312fc292a4e0e0b2bc9, disabling compactions & flushes 2024-12-06T07:53:15,698 DEBUG [RS:0;aa9080fe83d2:43895 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T07:53:15,698 INFO [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,698 DEBUG [RS:0;aa9080fe83d2:43895 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,698 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T07:53:15,698 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,698 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T07:53:15,698 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. after waiting 0 ms 2024-12-06T07:53:15,698 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-06T07:53:15,698 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T07:53:15,698 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,698 DEBUG [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1325): Online Regions={9bf2b886a5e12312fc292a4e0e0b2bc9=TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9.} 2024-12-06T07:53:15,698 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-06T07:53:15,698 DEBUG [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1351): Waiting on 9bf2b886a5e12312fc292a4e0e0b2bc9 2024-12-06T07:53:15,698 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,698 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-06T07:53:15,698 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,699 DEBUG [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-06T07:53:15,699 DEBUG [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-06T07:53:15,699 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-06T07:53:15,699 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,699 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-06T07:53:15,699 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,699 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-06T07:53:15,699 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-06T07:53:15,699 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-06T07:53:15,699 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,699 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-06T07:53:15,702 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741835_1011 (size=93) 2024-12-06T07:53:15,703 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741835_1011 (size=93) 2024-12-06T07:53:15,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741835_1011 (size=93) 2024-12-06T07:53:15,706 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/default/TestHBaseWalOnEC/9bf2b886a5e12312fc292a4e0e0b2bc9/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-06T07:53:15,706 DEBUG [RS:1;aa9080fe83d2:34373 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/oldWALs 2024-12-06T07:53:15,706 INFO [RS:1;aa9080fe83d2:34373 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog aa9080fe83d2%2C34373%2C1733471593541:(num 1733471594207) 2024-12-06T07:53:15,706 DEBUG [RS:1;aa9080fe83d2:34373 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,706 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:15,706 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T07:53:15,707 INFO [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,706 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.ChoreService(370): Chore service for: regionserver/aa9080fe83d2:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T07:53:15,707 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 9bf2b886a5e12312fc292a4e0e0b2bc9: Waiting for close lock at 1733471595698Running coprocessor pre-close hooks at 1733471595698Disabling compacts and flushes for region at 1733471595698Disabling writes for close at 1733471595698Writing region close event to WAL at 1733471595699 (+1 ms)Running coprocessor post-close hooks at 1733471595706 (+7 ms)Closed at 1733471595706 2024-12-06T07:53:15,707 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T07:53:15,707 INFO [regionserver/aa9080fe83d2:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T07:53:15,707 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T07:53:15,707 DEBUG [RS_CLOSE_REGION-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9. 2024-12-06T07:53:15,707 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T07:53:15,707 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T07:53:15,707 INFO [RS:1;aa9080fe83d2:34373 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:34373 2024-12-06T07:53:15,710 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/aa9080fe83d2,34373,1733471593541 2024-12-06T07:53:15,710 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T07:53:15,710 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T07:53:15,710 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [aa9080fe83d2,34373,1733471593541] 2024-12-06T07:53:15,713 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/aa9080fe83d2,34373,1733471593541 already deleted, retry=false 2024-12-06T07:53:15,713 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; aa9080fe83d2,34373,1733471593541 expired; onlineServers=2 2024-12-06T07:53:15,721 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/info/da22777b1a9c4d9da86e6aa0e75aab57 is 153, key is TestHBaseWalOnEC,,1733471594713.9bf2b886a5e12312fc292a4e0e0b2bc9./info:regioninfo/1733471595092/Put/seqid=0 2024-12-06T07:53:15,727 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741840_1016 (size=6637) 2024-12-06T07:53:15,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741840_1016 (size=6637) 2024-12-06T07:53:15,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741840_1016 (size=6637) 2024-12-06T07:53:15,729 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/info/da22777b1a9c4d9da86e6aa0e75aab57 2024-12-06T07:53:15,752 INFO [regionserver/aa9080fe83d2:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:15,753 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/ns/4cd65912f95b4a3c845e48511d204dcb is 43, key is default/ns:d/1733471594629/Put/seqid=0 2024-12-06T07:53:15,759 INFO [regionserver/aa9080fe83d2:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:15,759 INFO [regionserver/aa9080fe83d2:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:15,760 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741841_1017 (size=5153) 2024-12-06T07:53:15,761 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741841_1017 (size=5153) 2024-12-06T07:53:15,761 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741841_1017 (size=5153) 2024-12-06T07:53:15,761 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/ns/4cd65912f95b4a3c845e48511d204dcb 2024-12-06T07:53:15,784 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/table/895ddf5d3e034ce4a35617f39e1792af is 52, key is TestHBaseWalOnEC/table:state/1733471595106/Put/seqid=0 2024-12-06T07:53:15,790 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741842_1018 (size=5249) 2024-12-06T07:53:15,791 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741842_1018 (size=5249) 2024-12-06T07:53:15,791 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741842_1018 (size=5249) 2024-12-06T07:53:15,792 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/table/895ddf5d3e034ce4a35617f39e1792af 2024-12-06T07:53:15,800 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/info/da22777b1a9c4d9da86e6aa0e75aab57 as hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/info/da22777b1a9c4d9da86e6aa0e75aab57 2024-12-06T07:53:15,807 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/info/da22777b1a9c4d9da86e6aa0e75aab57, entries=10, sequenceid=11, filesize=6.5 K 2024-12-06T07:53:15,808 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/ns/4cd65912f95b4a3c845e48511d204dcb as hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/ns/4cd65912f95b4a3c845e48511d204dcb 2024-12-06T07:53:15,811 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:15,811 INFO [RS:1;aa9080fe83d2:34373 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T07:53:15,812 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:34373-0x10186c34e6c0002, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:15,812 INFO [RS:1;aa9080fe83d2:34373 {}] regionserver.HRegionServer(1031): Exiting; stopping=aa9080fe83d2,34373,1733471593541; zookeeper connection closed. 2024-12-06T07:53:15,812 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5e8cff37 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5e8cff37 2024-12-06T07:53:15,815 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/ns/4cd65912f95b4a3c845e48511d204dcb, entries=2, sequenceid=11, filesize=5.0 K 2024-12-06T07:53:15,816 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/.tmp/table/895ddf5d3e034ce4a35617f39e1792af as hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/table/895ddf5d3e034ce4a35617f39e1792af 2024-12-06T07:53:15,823 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/table/895ddf5d3e034ce4a35617f39e1792af, entries=2, sequenceid=11, filesize=5.1 K 2024-12-06T07:53:15,825 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 126ms, sequenceid=11, compaction requested=false 2024-12-06T07:53:15,830 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-06T07:53:15,831 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-06T07:53:15,831 INFO [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-06T07:53:15,831 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733471595699Running coprocessor pre-close hooks at 1733471595699Disabling compacts and flushes for region at 1733471595699Disabling writes for close at 1733471595699Obtaining lock to block concurrent updates at 1733471595699Preparing flush snapshotting stores in 1588230740 at 1733471595699Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733471595700 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733471595701 (+1 ms)Flushing 1588230740/info: creating writer at 1733471595701Flushing 1588230740/info: appending metadata at 1733471595720 (+19 ms)Flushing 1588230740/info: closing flushed file at 1733471595720Flushing 1588230740/ns: creating writer at 1733471595736 (+16 ms)Flushing 1588230740/ns: appending metadata at 1733471595752 (+16 ms)Flushing 1588230740/ns: closing flushed file at 1733471595752Flushing 1588230740/table: creating writer at 1733471595768 (+16 ms)Flushing 1588230740/table: appending metadata at 1733471595783 (+15 ms)Flushing 1588230740/table: closing flushed file at 1733471595783Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6daecd6: reopening flushed file at 1733471595799 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@14c774: reopening flushed file at 1733471595807 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@45894815: reopening flushed file at 1733471595815 (+8 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 126ms, sequenceid=11, compaction requested=false at 1733471595825 (+10 ms)Writing region close event to WAL at 1733471595826 (+1 ms)Running coprocessor post-close hooks at 1733471595831 (+5 ms)Closed at 1733471595831 2024-12-06T07:53:15,832 DEBUG [RS_CLOSE_META-regionserver/aa9080fe83d2:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-06T07:53:15,898 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(976): stopping server aa9080fe83d2,43895,1733471593506; all regions closed. 2024-12-06T07:53:15,899 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(976): stopping server aa9080fe83d2,36189,1733471593573; all regions closed. 2024-12-06T07:53:15,899 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,899 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,899 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,899 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,899 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,900 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,900 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,900 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,900 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,900 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741834_1010 (size=1298) 2024-12-06T07:53:15,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741834_1010 (size=1298) 2024-12-06T07:53:15,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741834_1010 (size=1298) 2024-12-06T07:53:15,905 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741836_1012 (size=2751) 2024-12-06T07:53:15,905 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741836_1012 (size=2751) 2024-12-06T07:53:15,906 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741836_1012 (size=2751) 2024-12-06T07:53:15,908 DEBUG [RS:0;aa9080fe83d2:43895 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/oldWALs 2024-12-06T07:53:15,908 INFO [RS:0;aa9080fe83d2:43895 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog aa9080fe83d2%2C43895%2C1733471593506:(num 1733471594202) 2024-12-06T07:53:15,908 DEBUG [RS:0;aa9080fe83d2:43895 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,909 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:15,909 DEBUG [RS:2;aa9080fe83d2:36189 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/oldWALs 2024-12-06T07:53:15,909 INFO [RS:2;aa9080fe83d2:36189 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog aa9080fe83d2%2C36189%2C1733471593573.meta:.meta(num 1733471594555) 2024-12-06T07:53:15,909 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T07:53:15,909 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.ChoreService(370): Chore service for: regionserver/aa9080fe83d2:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T07:53:15,909 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,909 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T07:53:15,909 INFO [regionserver/aa9080fe83d2:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T07:53:15,909 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,909 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T07:53:15,909 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T07:53:15,909 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,909 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T07:53:15,909 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,910 INFO [RS:0;aa9080fe83d2:43895 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:43895 2024-12-06T07:53:15,910 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:15,912 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/aa9080fe83d2,43895,1733471593506 2024-12-06T07:53:15,912 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T07:53:15,912 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T07:53:15,913 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741833_1009 (size=93) 2024-12-06T07:53:15,913 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741833_1009 (size=93) 2024-12-06T07:53:15,913 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [aa9080fe83d2,43895,1733471593506] 2024-12-06T07:53:15,915 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741833_1009 (size=93) 2024-12-06T07:53:15,917 DEBUG [RS:2;aa9080fe83d2:36189 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/oldWALs 2024-12-06T07:53:15,917 INFO [RS:2;aa9080fe83d2:36189 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog aa9080fe83d2%2C36189%2C1733471593573:(num 1733471594197) 2024-12-06T07:53:15,917 DEBUG [RS:2;aa9080fe83d2:36189 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T07:53:15,917 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T07:53:15,917 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T07:53:15,917 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/aa9080fe83d2,43895,1733471593506 already deleted, retry=false 2024-12-06T07:53:15,917 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; aa9080fe83d2,43895,1733471593506 expired; onlineServers=1 2024-12-06T07:53:15,917 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.ChoreService(370): Chore service for: regionserver/aa9080fe83d2:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T07:53:15,918 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T07:53:15,918 INFO [regionserver/aa9080fe83d2:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T07:53:15,918 INFO [RS:2;aa9080fe83d2:36189 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36189 2024-12-06T07:53:15,922 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/aa9080fe83d2,36189,1733471593573 2024-12-06T07:53:15,922 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T07:53:15,922 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T07:53:15,923 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [aa9080fe83d2,36189,1733471593573] 2024-12-06T07:53:15,924 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/aa9080fe83d2,36189,1733471593573 already deleted, retry=false 2024-12-06T07:53:15,925 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; aa9080fe83d2,36189,1733471593573 expired; onlineServers=0 2024-12-06T07:53:15,925 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'aa9080fe83d2,38929,1733471593457' ***** 2024-12-06T07:53:15,925 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-06T07:53:15,925 INFO [M:0;aa9080fe83d2:38929 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T07:53:15,925 INFO [M:0;aa9080fe83d2:38929 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T07:53:15,925 DEBUG [M:0;aa9080fe83d2:38929 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-06T07:53:15,925 DEBUG [M:0;aa9080fe83d2:38929 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-06T07:53:15,925 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-06T07:53:15,925 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.large.0-1733471593923 {}] cleaner.HFileCleaner(306): Exit Thread[master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.large.0-1733471593923,5,FailOnTimeoutGroup] 2024-12-06T07:53:15,925 DEBUG [master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.small.0-1733471593923 {}] cleaner.HFileCleaner(306): Exit Thread[master/aa9080fe83d2:0:becomeActiveMaster-HFileCleaner.small.0-1733471593923,5,FailOnTimeoutGroup] 2024-12-06T07:53:15,925 INFO [M:0;aa9080fe83d2:38929 {}] hbase.ChoreService(370): Chore service for: master/aa9080fe83d2:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-06T07:53:15,925 INFO [M:0;aa9080fe83d2:38929 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T07:53:15,925 DEBUG [M:0;aa9080fe83d2:38929 {}] master.HMaster(1795): Stopping service threads 2024-12-06T07:53:15,926 INFO [M:0;aa9080fe83d2:38929 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-06T07:53:15,926 INFO [M:0;aa9080fe83d2:38929 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-06T07:53:15,926 INFO [M:0;aa9080fe83d2:38929 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-06T07:53:15,926 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-06T07:53:15,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-06T07:53:15,927 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T07:53:15,927 DEBUG [M:0;aa9080fe83d2:38929 {}] zookeeper.ZKUtil(347): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-06T07:53:15,927 WARN [M:0;aa9080fe83d2:38929 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-06T07:53:15,927 INFO [M:0;aa9080fe83d2:38929 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/.lastflushedseqids 2024-12-06T07:53:15,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741843_1019 (size=127) 2024-12-06T07:53:15,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741843_1019 (size=127) 2024-12-06T07:53:15,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741843_1019 (size=127) 2024-12-06T07:53:15,945 INFO [M:0;aa9080fe83d2:38929 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-06T07:53:15,945 INFO [M:0;aa9080fe83d2:38929 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-06T07:53:15,945 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-06T07:53:15,945 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:15,945 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:15,945 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-06T07:53:15,945 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:15,945 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.84 KB heapSize=34.13 KB 2024-12-06T07:53:15,964 DEBUG [M:0;aa9080fe83d2:38929 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/9346a336effc438db3fd8363d5e0187a is 82, key is hbase:meta,,1/info:regioninfo/1733471594613/Put/seqid=0 2024-12-06T07:53:15,975 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741844_1020 (size=5672) 2024-12-06T07:53:15,976 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741844_1020 (size=5672) 2024-12-06T07:53:15,977 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741844_1020 (size=5672) 2024-12-06T07:53:15,977 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/9346a336effc438db3fd8363d5e0187a 2024-12-06T07:53:16,010 DEBUG [M:0;aa9080fe83d2:38929 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed04bc8e881f4637a461780865de1f9c is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733471595112/Put/seqid=0 2024-12-06T07:53:16,017 INFO [RS:0;aa9080fe83d2:43895 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T07:53:16,017 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:16,017 INFO [RS:0;aa9080fe83d2:43895 {}] regionserver.HRegionServer(1031): Exiting; stopping=aa9080fe83d2,43895,1733471593506; zookeeper connection closed. 2024-12-06T07:53:16,017 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43895-0x10186c34e6c0001, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:16,018 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741845_1021 (size=6440) 2024-12-06T07:53:16,020 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741845_1021 (size=6440) 2024-12-06T07:53:16,020 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741845_1021 (size=6440) 2024-12-06T07:53:16,021 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.15 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed04bc8e881f4637a461780865de1f9c 2024-12-06T07:53:16,024 INFO [RS:2;aa9080fe83d2:36189 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T07:53:16,024 INFO [RS:2;aa9080fe83d2:36189 {}] regionserver.HRegionServer(1031): Exiting; stopping=aa9080fe83d2,36189,1733471593573; zookeeper connection closed. 2024-12-06T07:53:16,024 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:16,024 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36189-0x10186c34e6c0003, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:16,028 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@36172809 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@36172809 2024-12-06T07:53:16,030 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@49d54469 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@49d54469 2024-12-06T07:53:16,031 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-06T07:53:16,049 DEBUG [M:0;aa9080fe83d2:38929 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ae8e730c326f4ffa994f7066b71a53bf is 69, key is aa9080fe83d2,34373,1733471593541/rs:state/1733471594024/Put/seqid=0 2024-12-06T07:53:16,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741846_1022 (size=5294) 2024-12-06T07:53:16,058 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741846_1022 (size=5294) 2024-12-06T07:53:16,059 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741846_1022 (size=5294) 2024-12-06T07:53:16,059 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ae8e730c326f4ffa994f7066b71a53bf 2024-12-06T07:53:16,068 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/9346a336effc438db3fd8363d5e0187a as hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/9346a336effc438db3fd8363d5e0187a 2024-12-06T07:53:16,076 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/9346a336effc438db3fd8363d5e0187a, entries=8, sequenceid=72, filesize=5.5 K 2024-12-06T07:53:16,077 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed04bc8e881f4637a461780865de1f9c as hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ed04bc8e881f4637a461780865de1f9c 2024-12-06T07:53:16,084 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ed04bc8e881f4637a461780865de1f9c, entries=8, sequenceid=72, filesize=6.3 K 2024-12-06T07:53:16,085 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ae8e730c326f4ffa994f7066b71a53bf as hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/ae8e730c326f4ffa994f7066b71a53bf 2024-12-06T07:53:16,091 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:32919/user/jenkins/test-data/bbaca901-c685-518c-c99c-783b071b3dee/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/ae8e730c326f4ffa994f7066b71a53bf, entries=3, sequenceid=72, filesize=5.2 K 2024-12-06T07:53:16,093 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 148ms, sequenceid=72, compaction requested=false 2024-12-06T07:53:16,094 INFO [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T07:53:16,094 DEBUG [M:0;aa9080fe83d2:38929 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733471595945Disabling compacts and flushes for region at 1733471595945Disabling writes for close at 1733471595945Obtaining lock to block concurrent updates at 1733471595945Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733471595945Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27480, getHeapSize=34880, getOffHeapSize=0, getCellsCount=85 at 1733471595946 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733471595947 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733471595947Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733471595964 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733471595964Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733471595988 (+24 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733471596010 (+22 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733471596010Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733471596029 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733471596049 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733471596049Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4b9f75ee: reopening flushed file at 1733471596067 (+18 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@94d57b1: reopening flushed file at 1733471596076 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2f1c3b4a: reopening flushed file at 1733471596084 (+8 ms)Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 148ms, sequenceid=72, compaction requested=false at 1733471596093 (+9 ms)Writing region close event to WAL at 1733471596094 (+1 ms)Closed at 1733471596094 2024-12-06T07:53:16,095 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:16,095 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:16,095 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:16,095 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:16,095 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T07:53:16,097 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35709 is added to blk_1073741830_1006 (size=32683) 2024-12-06T07:53:16,098 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45937 is added to blk_1073741830_1006 (size=32683) 2024-12-06T07:53:16,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37709 is added to blk_1073741830_1006 (size=32683) 2024-12-06T07:53:16,099 INFO [M:0;aa9080fe83d2:38929 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-06T07:53:16,099 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T07:53:16,099 INFO [M:0;aa9080fe83d2:38929 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:38929 2024-12-06T07:53:16,100 INFO [M:0;aa9080fe83d2:38929 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T07:53:16,204 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:16,204 INFO [M:0;aa9080fe83d2:38929 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T07:53:16,204 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38929-0x10186c34e6c0000, quorum=127.0.0.1:63156, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T07:53:16,207 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@11ec8dac{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:16,207 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@582dea15{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T07:53:16,207 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T07:53:16,208 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@660b8bbc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T07:53:16,208 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@b7fa3ef{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir/,STOPPED} 2024-12-06T07:53:16,209 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T07:53:16,209 WARN [BP-88630231-172.17.0.2-1733471592549 heartbeating to localhost/127.0.0.1:32919 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T07:53:16,209 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T07:53:16,209 WARN [BP-88630231-172.17.0.2-1733471592549 heartbeating to localhost/127.0.0.1:32919 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-88630231-172.17.0.2-1733471592549 (Datanode Uuid 81a92cbd-4f68-4ced-86eb-04e3045863c0) service to localhost/127.0.0.1:32919 2024-12-06T07:53:16,210 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data5/current/BP-88630231-172.17.0.2-1733471592549 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:16,210 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data6/current/BP-88630231-172.17.0.2-1733471592549 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:16,211 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T07:53:16,213 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@65d24fa8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:16,213 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@335d6851{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T07:53:16,213 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T07:53:16,214 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@505d9ca3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T07:53:16,214 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@55791d09{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir/,STOPPED} 2024-12-06T07:53:16,215 WARN [BP-88630231-172.17.0.2-1733471592549 heartbeating to localhost/127.0.0.1:32919 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T07:53:16,215 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T07:53:16,216 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T07:53:16,216 WARN [BP-88630231-172.17.0.2-1733471592549 heartbeating to localhost/127.0.0.1:32919 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-88630231-172.17.0.2-1733471592549 (Datanode Uuid 0b4abb8a-02fe-4d81-a82b-604f7a271955) service to localhost/127.0.0.1:32919 2024-12-06T07:53:16,216 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data3/current/BP-88630231-172.17.0.2-1733471592549 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:16,216 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data4/current/BP-88630231-172.17.0.2-1733471592549 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:16,217 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T07:53:16,219 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@a49b909{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T07:53:16,219 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1c513200{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T07:53:16,219 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T07:53:16,220 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7c7a8992{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T07:53:16,220 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@78ab2b00{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir/,STOPPED} 2024-12-06T07:53:16,221 WARN [BP-88630231-172.17.0.2-1733471592549 heartbeating to localhost/127.0.0.1:32919 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T07:53:16,221 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T07:53:16,221 WARN [BP-88630231-172.17.0.2-1733471592549 heartbeating to localhost/127.0.0.1:32919 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-88630231-172.17.0.2-1733471592549 (Datanode Uuid dac1264a-8410-43a4-9005-5f20fb889398) service to localhost/127.0.0.1:32919 2024-12-06T07:53:16,221 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T07:53:16,222 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data1/current/BP-88630231-172.17.0.2-1733471592549 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:16,222 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/cluster_74187dd0-80b6-f04d-0ba2-714a868fa399/data/data2/current/BP-88630231-172.17.0.2-1733471592549 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T07:53:16,223 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T07:53:16,229 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4453029d{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-06T07:53:16,230 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5ad015a3{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T07:53:16,230 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T07:53:16,230 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@28bcd4bd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T07:53:16,230 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@41ab5cc4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/461b9302-dc9e-43b3-207c-e225cd381214/hadoop.log.dir/,STOPPED} 2024-12-06T07:53:16,238 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-06T07:53:16,263 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-06T07:53:16,271 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=145 (was 84) - Thread LEAK? -, OpenFileDescriptor=520 (was 447) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=212 (was 187) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=3559 (was 3819)