2024-12-12 01:43:20,357 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-12 01:43:20,368 main DEBUG Took 0.008732 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-12 01:43:20,368 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-12 01:43:20,369 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-12 01:43:20,369 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-12 01:43:20,370 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,379 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-12 01:43:20,394 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,395 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,395 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,396 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,396 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,396 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,397 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,397 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,398 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,398 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,398 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,399 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,399 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,399 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,400 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,400 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,400 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,401 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,401 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,401 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,401 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,402 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,402 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,402 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-12 01:43:20,403 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,403 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-12 01:43:20,404 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-12 01:43:20,406 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-12 01:43:20,407 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-12 01:43:20,408 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-12 01:43:20,409 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-12 01:43:20,409 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-12 01:43:20,416 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-12 01:43:20,419 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-12 01:43:20,420 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-12 01:43:20,421 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-12 01:43:20,421 main DEBUG createAppenders(={Console}) 2024-12-12 01:43:20,422 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-12 01:43:20,422 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-12 01:43:20,422 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-12 01:43:20,423 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-12 01:43:20,423 main DEBUG OutputStream closed 2024-12-12 01:43:20,423 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-12 01:43:20,423 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-12 01:43:20,423 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-12 01:43:20,484 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-12 01:43:20,485 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-12 01:43:20,486 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-12 01:43:20,487 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-12 01:43:20,488 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-12 01:43:20,488 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-12 01:43:20,488 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-12 01:43:20,488 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-12 01:43:20,489 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-12 01:43:20,489 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-12 01:43:20,489 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-12 01:43:20,489 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-12 01:43:20,490 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-12 01:43:20,490 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-12 01:43:20,490 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-12 01:43:20,490 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-12 01:43:20,491 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-12 01:43:20,491 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-12 01:43:20,493 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-12 01:43:20,494 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-12 01:43:20,494 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-12 01:43:20,494 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-12T01:43:20,508 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-12 01:43:20,511 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-12 01:43:20,511 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-12T01:43:20,753 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7 2024-12-12T01:43:20,777 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647, deleteOnExit=true 2024-12-12T01:43:20,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/test.cache.data in system properties and HBase conf 2024-12-12T01:43:20,778 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.tmp.dir in system properties and HBase conf 2024-12-12T01:43:20,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir in system properties and HBase conf 2024-12-12T01:43:20,779 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-12T01:43:20,780 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-12T01:43:20,780 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-12T01:43:20,859 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-12T01:43:20,932 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-12T01:43:20,936 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-12T01:43:20,936 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-12T01:43:20,936 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-12T01:43:20,937 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-12T01:43:20,937 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-12T01:43:20,938 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-12T01:43:20,938 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-12T01:43:20,938 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-12T01:43:20,939 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-12T01:43:20,939 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/nfs.dump.dir in system properties and HBase conf 2024-12-12T01:43:20,939 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/java.io.tmpdir in system properties and HBase conf 2024-12-12T01:43:20,940 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-12T01:43:20,940 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-12T01:43:20,940 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-12T01:43:21,874 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-12T01:43:21,934 INFO [Time-limited test {}] log.Log(170): Logging initialized @2133ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-12T01:43:21,992 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:22,044 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-12T01:43:22,061 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-12T01:43:22,061 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-12T01:43:22,062 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-12T01:43:22,073 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:22,076 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@77f7f078{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir/,AVAILABLE} 2024-12-12T01:43:22,077 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@e67f021{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-12T01:43:22,235 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@63603efd{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/java.io.tmpdir/jetty-localhost-34971-hadoop-hdfs-3_4_1-tests_jar-_-any-429609282963962161/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-12T01:43:22,242 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3f607d40{HTTP/1.1, (http/1.1)}{localhost:34971} 2024-12-12T01:43:22,242 INFO [Time-limited test {}] server.Server(415): Started @2442ms 2024-12-12T01:43:22,703 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:22,711 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-12T01:43:22,712 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-12T01:43:22,713 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-12T01:43:22,713 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-12T01:43:22,713 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2d15633c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir/,AVAILABLE} 2024-12-12T01:43:22,714 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@79f7513c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-12T01:43:22,824 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@103787ab{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/java.io.tmpdir/jetty-localhost-39425-hadoop-hdfs-3_4_1-tests_jar-_-any-11596745318475385790/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:22,824 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@23a1a692{HTTP/1.1, (http/1.1)}{localhost:39425} 2024-12-12T01:43:22,825 INFO [Time-limited test {}] server.Server(415): Started @3024ms 2024-12-12T01:43:22,880 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-12T01:43:22,995 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:22,999 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-12T01:43:23,002 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-12T01:43:23,003 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-12T01:43:23,003 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-12T01:43:23,003 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3f04a93b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir/,AVAILABLE} 2024-12-12T01:43:23,004 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1e752fdd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-12T01:43:23,105 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@43fe7c8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/java.io.tmpdir/jetty-localhost-37311-hadoop-hdfs-3_4_1-tests_jar-_-any-9494404202770006251/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:23,106 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@6fecef94{HTTP/1.1, (http/1.1)}{localhost:37311} 2024-12-12T01:43:23,106 INFO [Time-limited test {}] server.Server(415): Started @3306ms 2024-12-12T01:43:23,108 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-12T01:43:23,141 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:23,145 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-12T01:43:23,147 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-12T01:43:23,147 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-12T01:43:23,148 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-12T01:43:23,151 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@f4b982a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir/,AVAILABLE} 2024-12-12T01:43:23,152 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@36ef30b2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-12T01:43:23,242 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@793c26cb{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/java.io.tmpdir/jetty-localhost-38069-hadoop-hdfs-3_4_1-tests_jar-_-any-4284311667282155226/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:23,243 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3e0ab7c3{HTTP/1.1, (http/1.1)}{localhost:38069} 2024-12-12T01:43:23,243 INFO [Time-limited test {}] server.Server(415): Started @3443ms 2024-12-12T01:43:23,245 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-12T01:43:24,164 WARN [Thread-125 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data2/current/BP-986496953-172.17.0.2-1733967801420/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:24,164 WARN [Thread-122 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data3/current/BP-986496953-172.17.0.2-1733967801420/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:24,164 WARN [Thread-123 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data4/current/BP-986496953-172.17.0.2-1733967801420/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:24,164 WARN [Thread-124 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data1/current/BP-986496953-172.17.0.2-1733967801420/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:24,195 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-12T01:43:24,195 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-12T01:43:24,236 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x79f7db950ca2cf91 with lease ID 0x9d69e2ecf95c17b2: Processing first storage report for DS-235c4a3e-5193-48ae-bc3f-b780dd15a9e1 from datanode DatanodeRegistration(127.0.0.1:43843, datanodeUuid=03106b5e-0306-482f-b556-190a78e07bb5, infoPort=41617, infoSecurePort=0, ipcPort=43735, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420) 2024-12-12T01:43:24,237 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x79f7db950ca2cf91 with lease ID 0x9d69e2ecf95c17b2: from storage DS-235c4a3e-5193-48ae-bc3f-b780dd15a9e1 node DatanodeRegistration(127.0.0.1:43843, datanodeUuid=03106b5e-0306-482f-b556-190a78e07bb5, infoPort=41617, infoSecurePort=0, ipcPort=43735, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2024-12-12T01:43:24,237 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x99d22bfd82e4f31e with lease ID 0x9d69e2ecf95c17b1: Processing first storage report for DS-099cf96d-ba8f-4ba1-bf7a-8377684c6888 from datanode DatanodeRegistration(127.0.0.1:41121, datanodeUuid=132dc7f6-a5d8-460e-b826-15809fcc7046, infoPort=33243, infoSecurePort=0, ipcPort=39843, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420) 2024-12-12T01:43:24,237 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x99d22bfd82e4f31e with lease ID 0x9d69e2ecf95c17b1: from storage DS-099cf96d-ba8f-4ba1-bf7a-8377684c6888 node DatanodeRegistration(127.0.0.1:41121, datanodeUuid=132dc7f6-a5d8-460e-b826-15809fcc7046, infoPort=33243, infoSecurePort=0, ipcPort=39843, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:24,238 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x79f7db950ca2cf91 with lease ID 0x9d69e2ecf95c17b2: Processing first storage report for DS-76c184dd-d2b9-499e-919f-cc9d5c9cdca5 from datanode DatanodeRegistration(127.0.0.1:43843, datanodeUuid=03106b5e-0306-482f-b556-190a78e07bb5, infoPort=41617, infoSecurePort=0, ipcPort=43735, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420) 2024-12-12T01:43:24,238 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x79f7db950ca2cf91 with lease ID 0x9d69e2ecf95c17b2: from storage DS-76c184dd-d2b9-499e-919f-cc9d5c9cdca5 node DatanodeRegistration(127.0.0.1:43843, datanodeUuid=03106b5e-0306-482f-b556-190a78e07bb5, infoPort=41617, infoSecurePort=0, ipcPort=43735, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-12T01:43:24,238 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x99d22bfd82e4f31e with lease ID 0x9d69e2ecf95c17b1: Processing first storage report for DS-2e711ffa-5f95-4646-8e92-c0ba6ec3210d from datanode DatanodeRegistration(127.0.0.1:41121, datanodeUuid=132dc7f6-a5d8-460e-b826-15809fcc7046, infoPort=33243, infoSecurePort=0, ipcPort=39843, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420) 2024-12-12T01:43:24,238 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x99d22bfd82e4f31e with lease ID 0x9d69e2ecf95c17b1: from storage DS-2e711ffa-5f95-4646-8e92-c0ba6ec3210d node DatanodeRegistration(127.0.0.1:41121, datanodeUuid=132dc7f6-a5d8-460e-b826-15809fcc7046, infoPort=33243, infoSecurePort=0, ipcPort=39843, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:24,266 WARN [Thread-143 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data6/current/BP-986496953-172.17.0.2-1733967801420/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:24,266 WARN [Thread-142 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data5/current/BP-986496953-172.17.0.2-1733967801420/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:24,281 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-12T01:43:24,285 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xa4d144a293df5d24 with lease ID 0x9d69e2ecf95c17b3: Processing first storage report for DS-45721179-d16e-4313-afd5-39cae4e4e581 from datanode DatanodeRegistration(127.0.0.1:39973, datanodeUuid=5ec7a961-dbb5-4343-9349-880ff4c9a610, infoPort=38593, infoSecurePort=0, ipcPort=40179, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420) 2024-12-12T01:43:24,286 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa4d144a293df5d24 with lease ID 0x9d69e2ecf95c17b3: from storage DS-45721179-d16e-4313-afd5-39cae4e4e581 node DatanodeRegistration(127.0.0.1:39973, datanodeUuid=5ec7a961-dbb5-4343-9349-880ff4c9a610, infoPort=38593, infoSecurePort=0, ipcPort=40179, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:24,286 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xa4d144a293df5d24 with lease ID 0x9d69e2ecf95c17b3: Processing first storage report for DS-3da6d07f-d1bb-4b3a-a482-5e2c0a7f41be from datanode DatanodeRegistration(127.0.0.1:39973, datanodeUuid=5ec7a961-dbb5-4343-9349-880ff4c9a610, infoPort=38593, infoSecurePort=0, ipcPort=40179, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420) 2024-12-12T01:43:24,286 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa4d144a293df5d24 with lease ID 0x9d69e2ecf95c17b3: from storage DS-3da6d07f-d1bb-4b3a-a482-5e2c0a7f41be node DatanodeRegistration(127.0.0.1:39973, datanodeUuid=5ec7a961-dbb5-4343-9349-880ff4c9a610, infoPort=38593, infoSecurePort=0, ipcPort=40179, storageInfo=lv=-57;cid=testClusterID;nsid=495427955;c=1733967801420), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:24,354 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7 2024-12-12T01:43:24,428 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-12T01:43:24,475 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=157, OpenFileDescriptor=393, MaxFileDescriptor=1048576, SystemLoadAverage=349, ProcessCount=11, AvailableMemoryMB=9713 2024-12-12T01:43:24,477 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-12T01:43:24,484 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-12T01:43:24,564 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/zookeeper_0, clientPort=60989, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-12T01:43:24,574 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=60989 2024-12-12T01:43:24,582 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:24,584 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:24,651 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:24,652 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:24,694 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:45366 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:41121:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45366 dst: /127.0.0.1:41121 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:24,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-12T01:43:25,114 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:25,127 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38 with version=8 2024-12-12T01:43:25,128 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/hbase-staging 2024-12-12T01:43:25,207 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-12T01:43:25,444 INFO [Time-limited test {}] client.ConnectionUtils(128): master/b85a53667e07:0 server-side Connection retries=45 2024-12-12T01:43:25,452 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:25,453 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:25,457 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-12T01:43:25,457 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:25,457 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-12T01:43:25,571 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-12T01:43:25,625 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-12T01:43:25,635 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-12T01:43:25,640 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-12T01:43:25,662 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 6945 (auto-detected) 2024-12-12T01:43:25,663 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-12T01:43:25,678 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36419 2024-12-12T01:43:25,698 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:36419 connecting to ZooKeeper ensemble=127.0.0.1:60989 2024-12-12T01:43:25,801 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:364190x0, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-12T01:43:25,803 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:36419-0x10017beb7220000 connected 2024-12-12T01:43:25,879 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:25,885 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:25,899 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:25,902 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38, hbase.cluster.distributed=false 2024-12-12T01:43:25,922 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-12T01:43:25,926 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36419 2024-12-12T01:43:25,927 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36419 2024-12-12T01:43:25,927 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36419 2024-12-12T01:43:25,928 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36419 2024-12-12T01:43:25,929 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36419 2024-12-12T01:43:26,014 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/b85a53667e07:0 server-side Connection retries=45 2024-12-12T01:43:26,016 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,016 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,016 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-12T01:43:26,016 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,016 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-12T01:43:26,018 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-12T01:43:26,020 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-12T01:43:26,021 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:37801 2024-12-12T01:43:26,022 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:37801 connecting to ZooKeeper ensemble=127.0.0.1:60989 2024-12-12T01:43:26,023 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:26,025 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:26,044 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:378010x0, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-12T01:43:26,045 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37801-0x10017beb7220001 connected 2024-12-12T01:43:26,045 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:26,052 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-12T01:43:26,060 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-12T01:43:26,063 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-12T01:43:26,068 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-12T01:43:26,068 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37801 2024-12-12T01:43:26,069 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37801 2024-12-12T01:43:26,069 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37801 2024-12-12T01:43:26,070 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37801 2024-12-12T01:43:26,071 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37801 2024-12-12T01:43:26,087 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/b85a53667e07:0 server-side Connection retries=45 2024-12-12T01:43:26,087 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,087 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,088 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-12T01:43:26,088 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,088 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-12T01:43:26,088 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-12T01:43:26,089 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-12T01:43:26,089 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:46061 2024-12-12T01:43:26,091 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:46061 connecting to ZooKeeper ensemble=127.0.0.1:60989 2024-12-12T01:43:26,092 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:26,094 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:26,128 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:460610x0, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-12T01:43:26,129 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:46061-0x10017beb7220002 connected 2024-12-12T01:43:26,129 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:26,130 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-12T01:43:26,131 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-12T01:43:26,132 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-12T01:43:26,134 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-12T01:43:26,135 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46061 2024-12-12T01:43:26,136 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46061 2024-12-12T01:43:26,136 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46061 2024-12-12T01:43:26,137 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46061 2024-12-12T01:43:26,137 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46061 2024-12-12T01:43:26,152 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/b85a53667e07:0 server-side Connection retries=45 2024-12-12T01:43:26,152 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,152 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,152 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-12T01:43:26,152 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:26,153 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-12T01:43:26,153 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-12T01:43:26,153 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-12T01:43:26,154 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:37251 2024-12-12T01:43:26,155 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:37251 connecting to ZooKeeper ensemble=127.0.0.1:60989 2024-12-12T01:43:26,156 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:26,158 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:26,169 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:372510x0, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-12T01:43:26,170 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37251-0x10017beb7220003 connected 2024-12-12T01:43:26,170 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:372510x0, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:26,171 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-12T01:43:26,172 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-12T01:43:26,173 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-12T01:43:26,174 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-12T01:43:26,175 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37251 2024-12-12T01:43:26,175 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37251 2024-12-12T01:43:26,176 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37251 2024-12-12T01:43:26,179 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37251 2024-12-12T01:43:26,179 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37251 2024-12-12T01:43:26,192 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;b85a53667e07:36419 2024-12-12T01:43:26,193 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/b85a53667e07,36419,1733967805298 2024-12-12T01:43:26,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:26,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:26,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:26,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:26,205 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/b85a53667e07,36419,1733967805298 2024-12-12T01:43:26,227 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-12T01:43:26,227 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-12T01:43:26,227 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-12T01:43:26,227 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,227 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,227 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,228 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,228 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-12T01:43:26,229 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/b85a53667e07,36419,1733967805298 from backup master directory 2024-12-12T01:43:26,236 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/b85a53667e07,36419,1733967805298 2024-12-12T01:43:26,236 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:26,236 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:26,236 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:26,236 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:26,237 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-12T01:43:26,237 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=b85a53667e07,36419,1733967805298 2024-12-12T01:43:26,239 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-12T01:43:26,240 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-12T01:43:26,294 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/hbase.id] with ID: 88a58891-d473-4c2b-b7e3-34540e107cf8 2024-12-12T01:43:26,295 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/.tmp/hbase.id 2024-12-12T01:43:26,301 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:26,302 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:26,305 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:51080 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:43843:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51080 dst: /127.0.0.1:43843 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:26,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-12T01:43:26,312 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:26,312 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/.tmp/hbase.id]:[hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/hbase.id] 2024-12-12T01:43:26,354 INFO [master/b85a53667e07:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:26,358 INFO [master/b85a53667e07:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-12T01:43:26,374 INFO [master/b85a53667e07:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 15ms. 2024-12-12T01:43:26,411 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,411 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,411 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,411 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,426 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:26,426 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:26,429 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:45394 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:41121:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45394 dst: /127.0.0.1:41121 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:26,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-12T01:43:26,435 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:26,448 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-12T01:43:26,450 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-12T01:43:26,454 INFO [master/b85a53667e07:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-12T01:43:26,477 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:26,477 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:26,480 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:52714 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:39973:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52714 dst: /127.0.0.1:39973 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:26,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-12T01:43:26,486 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:26,501 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store 2024-12-12T01:43:26,518 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:26,518 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:26,521 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:52740 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:39973:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52740 dst: /127.0.0.1:39973 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:26,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-12T01:43:26,526 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:26,529 INFO [master/b85a53667e07:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-12T01:43:26,532 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:26,533 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-12T01:43:26,533 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:26,533 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:26,535 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-12T01:43:26,535 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:26,535 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:26,536 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733967806533Disabling compacts and flushes for region at 1733967806533Disabling writes for close at 1733967806535 (+2 ms)Writing region close event to WAL at 1733967806535Closed at 1733967806535 2024-12-12T01:43:26,538 WARN [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/.initializing 2024-12-12T01:43:26,538 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/WALs/b85a53667e07,36419,1733967805298 2024-12-12T01:43:26,545 INFO [master/b85a53667e07:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-12T01:43:26,559 INFO [master/b85a53667e07:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C36419%2C1733967805298, suffix=, logDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/WALs/b85a53667e07,36419,1733967805298, archiveDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/oldWALs, maxLogs=10 2024-12-12T01:43:26,583 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/WALs/b85a53667e07,36419,1733967805298/b85a53667e07%2C36419%2C1733967805298.1733967806562, exclude list is [], retry=0 2024-12-12T01:43:26,599 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:26,600 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43843,DS-235c4a3e-5193-48ae-bc3f-b780dd15a9e1,DISK] 2024-12-12T01:43:26,600 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39973,DS-45721179-d16e-4313-afd5-39cae4e4e581,DISK] 2024-12-12T01:43:26,600 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41121,DS-099cf96d-ba8f-4ba1-bf7a-8377684c6888,DISK] 2024-12-12T01:43:26,603 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-12T01:43:26,638 INFO [master/b85a53667e07:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/WALs/b85a53667e07,36419,1733967805298/b85a53667e07%2C36419%2C1733967805298.1733967806562 2024-12-12T01:43:26,639 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:38593:38593),(127.0.0.1/127.0.0.1:33243:33243),(127.0.0.1/127.0.0.1:41617:41617)] 2024-12-12T01:43:26,640 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-12T01:43:26,640 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:26,644 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,645 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,678 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,700 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-12T01:43:26,702 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:26,704 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:26,705 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,708 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-12T01:43:26,708 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:26,709 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-12T01:43:26,709 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,712 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-12T01:43:26,712 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:26,713 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-12T01:43:26,713 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,715 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-12T01:43:26,716 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:26,716 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-12T01:43:26,717 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,720 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,722 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,726 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,727 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,730 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-12T01:43:26,733 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:26,739 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-12T01:43:26,741 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71547504, jitterRate=0.06614089012145996}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-12T01:43:26,746 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733967806657Initializing all the Stores at 1733967806659 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967806659Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967806660 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967806660Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967806660Cleaning up temporary data from old regions at 1733967806727 (+67 ms)Region opened successfully at 1733967806746 (+19 ms) 2024-12-12T01:43:26,747 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-12T01:43:26,774 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1a8b2140, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=b85a53667e07/172.17.0.2:0 2024-12-12T01:43:26,798 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-12T01:43:26,807 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-12T01:43:26,807 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-12T01:43:26,809 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-12T01:43:26,810 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-12T01:43:26,814 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 3 msec 2024-12-12T01:43:26,814 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-12T01:43:26,834 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-12T01:43:26,841 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-12T01:43:26,869 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-12T01:43:26,873 INFO [master/b85a53667e07:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-12T01:43:26,875 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-12T01:43:26,886 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-12T01:43:26,888 INFO [master/b85a53667e07:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-12T01:43:26,891 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-12T01:43:26,902 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-12T01:43:26,904 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-12T01:43:26,910 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-12T01:43:26,931 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-12T01:43:26,935 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-12T01:43:26,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:26,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:26,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:26,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:26,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,945 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,947 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=b85a53667e07,36419,1733967805298, sessionid=0x10017beb7220000, setting cluster-up flag (Was=false) 2024-12-12T01:43:26,969 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,969 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,969 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,969 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:26,994 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-12T01:43:26,999 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=b85a53667e07,36419,1733967805298 2024-12-12T01:43:27,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:27,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:27,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:27,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:27,044 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-12T01:43:27,046 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=b85a53667e07,36419,1733967805298 2024-12-12T01:43:27,054 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-12T01:43:27,083 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(746): ClusterId : 88a58891-d473-4c2b-b7e3-34540e107cf8 2024-12-12T01:43:27,083 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(746): ClusterId : 88a58891-d473-4c2b-b7e3-34540e107cf8 2024-12-12T01:43:27,083 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(746): ClusterId : 88a58891-d473-4c2b-b7e3-34540e107cf8 2024-12-12T01:43:27,085 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-12T01:43:27,085 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-12T01:43:27,085 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-12T01:43:27,096 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-12T01:43:27,096 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-12T01:43:27,096 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-12T01:43:27,096 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-12T01:43:27,096 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-12T01:43:27,096 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-12T01:43:27,103 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-12T01:43:27,103 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-12T01:43:27,103 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-12T01:43:27,104 DEBUG [RS:0;b85a53667e07:37801 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@36132d1b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=b85a53667e07/172.17.0.2:0 2024-12-12T01:43:27,104 DEBUG [RS:1;b85a53667e07:46061 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4bab13de, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=b85a53667e07/172.17.0.2:0 2024-12-12T01:43:27,104 DEBUG [RS:2;b85a53667e07:37251 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1b2e4a73, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=b85a53667e07/172.17.0.2:0 2024-12-12T01:43:27,118 DEBUG [RS:1;b85a53667e07:46061 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;b85a53667e07:46061 2024-12-12T01:43:27,121 INFO [RS:1;b85a53667e07:46061 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-12T01:43:27,121 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;b85a53667e07:37251 2024-12-12T01:43:27,121 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;b85a53667e07:37801 2024-12-12T01:43:27,121 INFO [RS:1;b85a53667e07:46061 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-12T01:43:27,121 INFO [RS:2;b85a53667e07:37251 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-12T01:43:27,121 DEBUG [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-12T01:43:27,121 INFO [RS:0;b85a53667e07:37801 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-12T01:43:27,121 INFO [RS:2;b85a53667e07:37251 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-12T01:43:27,121 INFO [RS:0;b85a53667e07:37801 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-12T01:43:27,121 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-12T01:43:27,121 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-12T01:43:27,124 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,36419,1733967805298 with port=37251, startcode=1733967806152 2024-12-12T01:43:27,124 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,36419,1733967805298 with port=46061, startcode=1733967806086 2024-12-12T01:43:27,124 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,36419,1733967805298 with port=37801, startcode=1733967805985 2024-12-12T01:43:27,126 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-12T01:43:27,134 INFO [master/b85a53667e07:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-12T01:43:27,135 DEBUG [RS:1;b85a53667e07:46061 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-12T01:43:27,135 DEBUG [RS:2;b85a53667e07:37251 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-12T01:43:27,135 DEBUG [RS:0;b85a53667e07:37801 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-12T01:43:27,140 INFO [master/b85a53667e07:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-12T01:43:27,145 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: b85a53667e07,36419,1733967805298 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-12T01:43:27,153 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/b85a53667e07:0, corePoolSize=5, maxPoolSize=5 2024-12-12T01:43:27,153 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/b85a53667e07:0, corePoolSize=5, maxPoolSize=5 2024-12-12T01:43:27,153 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/b85a53667e07:0, corePoolSize=5, maxPoolSize=5 2024-12-12T01:43:27,154 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/b85a53667e07:0, corePoolSize=5, maxPoolSize=5 2024-12-12T01:43:27,154 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/b85a53667e07:0, corePoolSize=10, maxPoolSize=10 2024-12-12T01:43:27,154 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,154 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/b85a53667e07:0, corePoolSize=2, maxPoolSize=2 2024-12-12T01:43:27,154 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,160 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733967837160 2024-12-12T01:43:27,161 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-12T01:43:27,162 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-12T01:43:27,165 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-12T01:43:27,166 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-12T01:43:27,166 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-12T01:43:27,166 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-12T01:43:27,166 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-12T01:43:27,166 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-12T01:43:27,167 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,170 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50205, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-12T01:43:27,170 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52939, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-12T01:43:27,170 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:57023, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-12T01:43:27,171 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-12T01:43:27,172 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:27,172 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-12T01:43:27,173 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-12T01:43:27,173 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-12T01:43:27,176 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-12T01:43:27,179 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-12T01:43:27,179 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-12T01:43:27,181 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-12T01:43:27,182 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3334) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:667) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:4.0.0-alpha-1-SNAPSHOT] 2024-12-12T01:43:27,182 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.large.0-1733967807181,5,FailOnTimeoutGroup] 2024-12-12T01:43:27,183 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.small.0-1733967807182,5,FailOnTimeoutGroup] 2024-12-12T01:43:27,183 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,183 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-12T01:43:27,185 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:27,185 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,185 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:27,185 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,189 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:52768 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:39973:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52768 dst: /127.0.0.1:39973 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:27,197 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-12T01:43:27,197 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:27,199 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-12T01:43:27,199 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38 2024-12-12T01:43:27,201 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-12T01:43:27,201 DEBUG [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-12T01:43:27,201 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(2683): Master is not running yet 2024-12-12T01:43:27,201 WARN [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-12T01:43:27,201 WARN [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-12T01:43:27,201 WARN [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(841): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-12T01:43:27,206 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:27,207 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:27,211 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:45420 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:41121:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45420 dst: /127.0.0.1:41121 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:27,220 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-12T01:43:27,222 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:27,223 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:27,226 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-12T01:43:27,228 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-12T01:43:27,228 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:27,229 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:27,230 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-12T01:43:27,232 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-12T01:43:27,232 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:27,233 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:27,233 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-12T01:43:27,236 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-12T01:43:27,236 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:27,237 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:27,237 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-12T01:43:27,240 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-12T01:43:27,240 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:27,241 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:27,241 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-12T01:43:27,243 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740 2024-12-12T01:43:27,244 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740 2024-12-12T01:43:27,247 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-12T01:43:27,248 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-12T01:43:27,249 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-12T01:43:27,252 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-12T01:43:27,260 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-12T01:43:27,261 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64513808, jitterRate=-0.03866934776306152}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-12T01:43:27,264 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733967807223Initializing all the Stores at 1733967807225 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967807225Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967807225Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967807226 (+1 ms)Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967807226Cleaning up temporary data from old regions at 1733967807248 (+22 ms)Region opened successfully at 1733967807263 (+15 ms) 2024-12-12T01:43:27,264 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-12T01:43:27,264 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-12T01:43:27,264 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-12T01:43:27,264 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-12T01:43:27,264 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-12T01:43:27,266 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-12T01:43:27,266 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733967807264Disabling compacts and flushes for region at 1733967807264Disabling writes for close at 1733967807264Writing region close event to WAL at 1733967807265 (+1 ms)Closed at 1733967807266 (+1 ms) 2024-12-12T01:43:27,270 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-12T01:43:27,271 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-12T01:43:27,277 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-12T01:43:27,286 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-12T01:43:27,289 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-12T01:43:27,302 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,36419,1733967805298 with port=37801, startcode=1733967805985 2024-12-12T01:43:27,302 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,36419,1733967805298 with port=37251, startcode=1733967806152 2024-12-12T01:43:27,302 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,36419,1733967805298 with port=46061, startcode=1733967806086 2024-12-12T01:43:27,304 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] master.ServerManager(363): Checking decommissioned status of RegionServer b85a53667e07,37801,1733967805985 2024-12-12T01:43:27,307 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] master.ServerManager(517): Registering regionserver=b85a53667e07,37801,1733967805985 2024-12-12T01:43:27,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-12T01:43:27,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-12T01:43:27,314 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] master.ServerManager(363): Checking decommissioned status of RegionServer b85a53667e07,46061,1733967806086 2024-12-12T01:43:27,314 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] master.ServerManager(517): Registering regionserver=b85a53667e07,46061,1733967806086 2024-12-12T01:43:27,315 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38 2024-12-12T01:43:27,315 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40607 2024-12-12T01:43:27,315 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-12T01:43:27,317 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] master.ServerManager(363): Checking decommissioned status of RegionServer b85a53667e07,37251,1733967806152 2024-12-12T01:43:27,317 DEBUG [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38 2024-12-12T01:43:27,317 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36419 {}] master.ServerManager(517): Registering regionserver=b85a53667e07,37251,1733967806152 2024-12-12T01:43:27,317 DEBUG [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40607 2024-12-12T01:43:27,317 DEBUG [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-12T01:43:27,320 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38 2024-12-12T01:43:27,320 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40607 2024-12-12T01:43:27,320 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-12T01:43:27,379 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-12T01:43:27,394 DEBUG [RS:1;b85a53667e07:46061 {}] zookeeper.ZKUtil(111): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/b85a53667e07,46061,1733967806086 2024-12-12T01:43:27,394 DEBUG [RS:0;b85a53667e07:37801 {}] zookeeper.ZKUtil(111): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/b85a53667e07,37801,1733967805985 2024-12-12T01:43:27,395 WARN [RS:1;b85a53667e07:46061 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-12T01:43:27,395 WARN [RS:0;b85a53667e07:37801 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-12T01:43:27,395 INFO [RS:1;b85a53667e07:46061 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-12T01:43:27,395 INFO [RS:0;b85a53667e07:37801 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-12T01:43:27,395 DEBUG [RS:2;b85a53667e07:37251 {}] zookeeper.ZKUtil(111): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/b85a53667e07,37251,1733967806152 2024-12-12T01:43:27,395 WARN [RS:2;b85a53667e07:37251 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-12T01:43:27,395 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37801,1733967805985 2024-12-12T01:43:27,395 DEBUG [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,46061,1733967806086 2024-12-12T01:43:27,396 INFO [RS:2;b85a53667e07:37251 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-12T01:43:27,396 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37251,1733967806152 2024-12-12T01:43:27,397 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [b85a53667e07,37801,1733967805985] 2024-12-12T01:43:27,397 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [b85a53667e07,37251,1733967806152] 2024-12-12T01:43:27,397 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [b85a53667e07,46061,1733967806086] 2024-12-12T01:43:27,422 INFO [RS:0;b85a53667e07:37801 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-12T01:43:27,422 INFO [RS:1;b85a53667e07:46061 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-12T01:43:27,422 INFO [RS:2;b85a53667e07:37251 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-12T01:43:27,435 INFO [RS:1;b85a53667e07:46061 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-12T01:43:27,435 INFO [RS:0;b85a53667e07:37801 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-12T01:43:27,435 INFO [RS:2;b85a53667e07:37251 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-12T01:43:27,440 INFO [RS:0;b85a53667e07:37801 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-12T01:43:27,440 WARN [b85a53667e07:36419 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-12T01:43:27,440 INFO [RS:2;b85a53667e07:37251 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-12T01:43:27,440 INFO [RS:1;b85a53667e07:46061 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-12T01:43:27,440 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,440 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,440 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,441 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-12T01:43:27,442 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-12T01:43:27,443 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-12T01:43:27,446 INFO [RS:1;b85a53667e07:46061 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-12T01:43:27,446 INFO [RS:0;b85a53667e07:37801 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-12T01:43:27,446 INFO [RS:2;b85a53667e07:37251 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-12T01:43:27,448 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,448 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,448 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,448 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,448 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/b85a53667e07:0, corePoolSize=2, maxPoolSize=2 2024-12-12T01:43:27,449 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/b85a53667e07:0, corePoolSize=2, maxPoolSize=2 2024-12-12T01:43:27,449 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/b85a53667e07:0, corePoolSize=2, maxPoolSize=2 2024-12-12T01:43:27,449 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:27,449 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:27,449 DEBUG [RS:0;b85a53667e07:37801 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:27,450 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:27,450 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:27,450 DEBUG [RS:2;b85a53667e07:37251 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:27,450 DEBUG [RS:1;b85a53667e07:46061 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:27,451 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,37251,1733967806152-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,46061,1733967806086-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-12T01:43:27,451 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,37801,1733967805985-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-12T01:43:27,467 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HeapMemoryManager(213): Starting, tuneOn=false 2024-12-12T01:43:27,467 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HeapMemoryManager(213): Starting, tuneOn=false 2024-12-12T01:43:27,469 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,46061,1733967806086-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,469 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,37801,1733967805985-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,469 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,469 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,469 INFO [RS:0;b85a53667e07:37801 {}] regionserver.Replication(171): b85a53667e07,37801,1733967805985 started 2024-12-12T01:43:27,469 INFO [RS:1;b85a53667e07:46061 {}] regionserver.Replication(171): b85a53667e07,46061,1733967806086 started 2024-12-12T01:43:27,470 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HeapMemoryManager(213): Starting, tuneOn=false 2024-12-12T01:43:27,470 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,37251,1733967806152-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,470 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,470 INFO [RS:2;b85a53667e07:37251 {}] regionserver.Replication(171): b85a53667e07,37251,1733967806152 started 2024-12-12T01:43:27,485 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,485 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,486 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1482): Serving as b85a53667e07,37801,1733967805985, RpcServer on b85a53667e07/172.17.0.2:37801, sessionid=0x10017beb7220001 2024-12-12T01:43:27,486 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(1482): Serving as b85a53667e07,46061,1733967806086, RpcServer on b85a53667e07/172.17.0.2:46061, sessionid=0x10017beb7220002 2024-12-12T01:43:27,486 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-12T01:43:27,486 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-12T01:43:27,487 DEBUG [RS:0;b85a53667e07:37801 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager b85a53667e07,37801,1733967805985 2024-12-12T01:43:27,487 DEBUG [RS:1;b85a53667e07:46061 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager b85a53667e07,46061,1733967806086 2024-12-12T01:43:27,487 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,37801,1733967805985' 2024-12-12T01:43:27,487 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,46061,1733967806086' 2024-12-12T01:43:27,487 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-12T01:43:27,487 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-12T01:43:27,488 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-12T01:43:27,488 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-12T01:43:27,488 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-12T01:43:27,488 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-12T01:43:27,488 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-12T01:43:27,488 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-12T01:43:27,489 DEBUG [RS:1;b85a53667e07:46061 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager b85a53667e07,46061,1733967806086 2024-12-12T01:43:27,489 DEBUG [RS:0;b85a53667e07:37801 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager b85a53667e07,37801,1733967805985 2024-12-12T01:43:27,489 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,46061,1733967806086' 2024-12-12T01:43:27,489 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,37801,1733967805985' 2024-12-12T01:43:27,489 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-12T01:43:27,489 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-12T01:43:27,489 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-12T01:43:27,489 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-12T01:43:27,490 DEBUG [RS:1;b85a53667e07:46061 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-12T01:43:27,490 INFO [RS:1;b85a53667e07:46061 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-12T01:43:27,490 INFO [RS:1;b85a53667e07:46061 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-12T01:43:27,490 DEBUG [RS:0;b85a53667e07:37801 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-12T01:43:27,490 INFO [RS:0;b85a53667e07:37801 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-12T01:43:27,490 INFO [RS:0;b85a53667e07:37801 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-12T01:43:27,491 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:27,491 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1482): Serving as b85a53667e07,37251,1733967806152, RpcServer on b85a53667e07/172.17.0.2:37251, sessionid=0x10017beb7220003 2024-12-12T01:43:27,491 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-12T01:43:27,491 DEBUG [RS:2;b85a53667e07:37251 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager b85a53667e07,37251,1733967806152 2024-12-12T01:43:27,491 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,37251,1733967806152' 2024-12-12T01:43:27,492 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-12T01:43:27,492 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-12T01:43:27,493 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-12T01:43:27,493 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-12T01:43:27,493 DEBUG [RS:2;b85a53667e07:37251 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager b85a53667e07,37251,1733967806152 2024-12-12T01:43:27,493 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,37251,1733967806152' 2024-12-12T01:43:27,493 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-12T01:43:27,494 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-12T01:43:27,494 DEBUG [RS:2;b85a53667e07:37251 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-12T01:43:27,495 INFO [RS:2;b85a53667e07:37251 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-12T01:43:27,495 INFO [RS:2;b85a53667e07:37251 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-12T01:43:27,596 INFO [RS:2;b85a53667e07:37251 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-12T01:43:27,596 INFO [RS:1;b85a53667e07:46061 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-12T01:43:27,597 INFO [RS:0;b85a53667e07:37801 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-12T01:43:27,602 INFO [RS:2;b85a53667e07:37251 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C37251%2C1733967806152, suffix=, logDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37251,1733967806152, archiveDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/oldWALs, maxLogs=32 2024-12-12T01:43:27,602 INFO [RS:0;b85a53667e07:37801 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C37801%2C1733967805985, suffix=, logDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37801,1733967805985, archiveDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/oldWALs, maxLogs=32 2024-12-12T01:43:27,602 INFO [RS:1;b85a53667e07:46061 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C46061%2C1733967806086, suffix=, logDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,46061,1733967806086, archiveDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/oldWALs, maxLogs=32 2024-12-12T01:43:27,616 DEBUG [RS:2;b85a53667e07:37251 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37251,1733967806152/b85a53667e07%2C37251%2C1733967806152.1733967807605, exclude list is [], retry=0 2024-12-12T01:43:27,618 DEBUG [RS:1;b85a53667e07:46061 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,46061,1733967806086/b85a53667e07%2C46061%2C1733967806086.1733967807605, exclude list is [], retry=0 2024-12-12T01:43:27,621 DEBUG [RS:0;b85a53667e07:37801 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37801,1733967805985/b85a53667e07%2C37801%2C1733967805985.1733967807605, exclude list is [], retry=0 2024-12-12T01:43:27,622 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39973,DS-45721179-d16e-4313-afd5-39cae4e4e581,DISK] 2024-12-12T01:43:27,622 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39973,DS-45721179-d16e-4313-afd5-39cae4e4e581,DISK] 2024-12-12T01:43:27,622 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43843,DS-235c4a3e-5193-48ae-bc3f-b780dd15a9e1,DISK] 2024-12-12T01:43:27,622 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41121,DS-099cf96d-ba8f-4ba1-bf7a-8377684c6888,DISK] 2024-12-12T01:43:27,622 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41121,DS-099cf96d-ba8f-4ba1-bf7a-8377684c6888,DISK] 2024-12-12T01:43:27,623 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43843,DS-235c4a3e-5193-48ae-bc3f-b780dd15a9e1,DISK] 2024-12-12T01:43:27,657 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41121,DS-099cf96d-ba8f-4ba1-bf7a-8377684c6888,DISK] 2024-12-12T01:43:27,657 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39973,DS-45721179-d16e-4313-afd5-39cae4e4e581,DISK] 2024-12-12T01:43:27,657 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43843,DS-235c4a3e-5193-48ae-bc3f-b780dd15a9e1,DISK] 2024-12-12T01:43:27,662 INFO [RS:1;b85a53667e07:46061 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,46061,1733967806086/b85a53667e07%2C46061%2C1733967806086.1733967807605 2024-12-12T01:43:27,662 INFO [RS:2;b85a53667e07:37251 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37251,1733967806152/b85a53667e07%2C37251%2C1733967806152.1733967807605 2024-12-12T01:43:27,663 DEBUG [RS:1;b85a53667e07:46061 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:38593:38593),(127.0.0.1/127.0.0.1:41617:41617),(127.0.0.1/127.0.0.1:33243:33243)] 2024-12-12T01:43:27,663 DEBUG [RS:2;b85a53667e07:37251 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:38593:38593),(127.0.0.1/127.0.0.1:33243:33243),(127.0.0.1/127.0.0.1:41617:41617)] 2024-12-12T01:43:27,668 INFO [RS:0;b85a53667e07:37801 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37801,1733967805985/b85a53667e07%2C37801%2C1733967805985.1733967807605 2024-12-12T01:43:27,668 DEBUG [RS:0;b85a53667e07:37801 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:33243:33243),(127.0.0.1/127.0.0.1:41617:41617),(127.0.0.1/127.0.0.1:38593:38593)] 2024-12-12T01:43:27,692 DEBUG [b85a53667e07:36419 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-12T01:43:27,702 DEBUG [b85a53667e07:36419 {}] balancer.BalancerClusterState(204): Hosts are {b85a53667e07=0} racks are {/default-rack=0} 2024-12-12T01:43:27,708 DEBUG [b85a53667e07:36419 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-12T01:43:27,708 DEBUG [b85a53667e07:36419 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-12T01:43:27,708 DEBUG [b85a53667e07:36419 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-12T01:43:27,708 DEBUG [b85a53667e07:36419 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-12T01:43:27,708 DEBUG [b85a53667e07:36419 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-12T01:43:27,708 DEBUG [b85a53667e07:36419 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-12T01:43:27,708 INFO [b85a53667e07:36419 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-12T01:43:27,708 INFO [b85a53667e07:36419 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-12T01:43:27,708 INFO [b85a53667e07:36419 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-12T01:43:27,708 DEBUG [b85a53667e07:36419 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-12T01:43:27,714 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=b85a53667e07,37801,1733967805985 2024-12-12T01:43:27,721 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as b85a53667e07,37801,1733967805985, state=OPENING 2024-12-12T01:43:27,752 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-12T01:43:27,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:27,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:27,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:27,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:27,804 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:27,804 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:27,804 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:27,804 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:27,807 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-12T01:43:27,810 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=b85a53667e07,37801,1733967805985}] 2024-12-12T01:43:28,009 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-12T01:43:28,011 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35317, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-12T01:43:28,022 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-12T01:43:28,023 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-12T01:43:28,023 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-12T01:43:28,026 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C37801%2C1733967805985.meta, suffix=.meta, logDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37801,1733967805985, archiveDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/oldWALs, maxLogs=32 2024-12-12T01:43:28,042 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37801,1733967805985/b85a53667e07%2C37801%2C1733967805985.meta.1733967808028.meta, exclude list is [], retry=0 2024-12-12T01:43:28,046 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41121,DS-099cf96d-ba8f-4ba1-bf7a-8377684c6888,DISK] 2024-12-12T01:43:28,046 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39973,DS-45721179-d16e-4313-afd5-39cae4e4e581,DISK] 2024-12-12T01:43:28,046 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43843,DS-235c4a3e-5193-48ae-bc3f-b780dd15a9e1,DISK] 2024-12-12T01:43:28,049 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/WALs/b85a53667e07,37801,1733967805985/b85a53667e07%2C37801%2C1733967805985.meta.1733967808028.meta 2024-12-12T01:43:28,049 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:33243:33243),(127.0.0.1/127.0.0.1:38593:38593),(127.0.0.1/127.0.0.1:41617:41617)] 2024-12-12T01:43:28,050 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-12T01:43:28,051 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-12T01:43:28,054 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-12T01:43:28,058 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-12T01:43:28,061 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-12T01:43:28,061 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:28,061 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-12T01:43:28,061 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-12T01:43:28,064 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-12T01:43:28,065 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-12T01:43:28,065 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:28,066 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:28,066 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-12T01:43:28,067 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-12T01:43:28,067 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:28,068 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:28,068 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-12T01:43:28,069 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-12T01:43:28,069 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:28,070 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:28,070 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-12T01:43:28,072 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-12T01:43:28,072 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:28,072 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:28,072 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-12T01:43:28,074 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740 2024-12-12T01:43:28,076 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740 2024-12-12T01:43:28,078 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-12T01:43:28,078 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-12T01:43:28,078 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-12T01:43:28,080 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-12T01:43:28,082 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74312969, jitterRate=0.10734952986240387}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-12T01:43:28,082 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-12T01:43:28,083 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733967808062Writing region info on filesystem at 1733967808062Initializing all the Stores at 1733967808063 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967808063Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967808064 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967808064Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967808064Cleaning up temporary data from old regions at 1733967808078 (+14 ms)Running coprocessor post-open hooks at 1733967808082 (+4 ms)Region opened successfully at 1733967808083 (+1 ms) 2024-12-12T01:43:28,088 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733967807981 2024-12-12T01:43:28,102 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-12T01:43:28,103 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-12T01:43:28,104 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=b85a53667e07,37801,1733967805985 2024-12-12T01:43:28,106 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as b85a53667e07,37801,1733967805985, state=OPEN 2024-12-12T01:43:28,135 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-12T01:43:28,135 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-12T01:43:28,135 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-12T01:43:28,135 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-12T01:43:28,136 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:28,136 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:28,136 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:28,136 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:28,136 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=b85a53667e07,37801,1733967805985 2024-12-12T01:43:28,144 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-12T01:43:28,144 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=b85a53667e07,37801,1733967805985 in 327 msec 2024-12-12T01:43:28,152 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-12T01:43:28,152 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 869 msec 2024-12-12T01:43:28,154 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-12T01:43:28,154 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-12T01:43:28,170 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-12T01:43:28,171 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=b85a53667e07,37801,1733967805985, seqNum=-1] 2024-12-12T01:43:28,187 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-12T01:43:28,189 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52525, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-12T01:43:28,207 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1240 sec 2024-12-12T01:43:28,207 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733967808207, completionTime=-1 2024-12-12T01:43:28,209 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-12T01:43:28,210 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-12T01:43:28,230 INFO [master/b85a53667e07:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-12T01:43:28,230 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733967868230 2024-12-12T01:43:28,230 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733967928230 2024-12-12T01:43:28,230 INFO [master/b85a53667e07:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 20 msec 2024-12-12T01:43:28,231 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-12T01:43:28,237 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,36419,1733967805298-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:28,237 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,36419,1733967805298-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:28,237 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,36419,1733967805298-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:28,239 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-b85a53667e07:36419, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:28,239 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:28,240 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:28,245 DEBUG [master/b85a53667e07:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-12T01:43:28,263 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.026sec 2024-12-12T01:43:28,265 INFO [master/b85a53667e07:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-12T01:43:28,265 INFO [master/b85a53667e07:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-12T01:43:28,266 INFO [master/b85a53667e07:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-12T01:43:28,266 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-12T01:43:28,266 INFO [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-12T01:43:28,267 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,36419,1733967805298-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-12T01:43:28,267 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,36419,1733967805298-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-12T01:43:28,271 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-12T01:43:28,272 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-12T01:43:28,272 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,36419,1733967805298-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:28,292 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@42f6fa1e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-12T01:43:28,296 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-12T01:43:28,296 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-12T01:43:28,299 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request b85a53667e07,36419,-1 for getting cluster id 2024-12-12T01:43:28,300 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-12T01:43:28,307 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '88a58891-d473-4c2b-b7e3-34540e107cf8' 2024-12-12T01:43:28,309 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-12T01:43:28,309 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "88a58891-d473-4c2b-b7e3-34540e107cf8" 2024-12-12T01:43:28,310 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@656aeb8b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-12T01:43:28,310 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [b85a53667e07,36419,-1] 2024-12-12T01:43:28,312 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-12T01:43:28,314 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:28,315 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49486, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-12T01:43:28,317 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4df1599, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-12T01:43:28,318 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-12T01:43:28,324 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=b85a53667e07,37801,1733967805985, seqNum=-1] 2024-12-12T01:43:28,325 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-12T01:43:28,327 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58700, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-12T01:43:28,343 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=b85a53667e07,36419,1733967805298 2024-12-12T01:43:28,347 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-12T01:43:28,350 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is b85a53667e07,36419,1733967805298 2024-12-12T01:43:28,352 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@45b7b835 2024-12-12T01:43:28,353 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-12T01:43:28,356 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49492, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-12T01:43:28,360 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-12T01:43:28,367 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-12T01:43:28,370 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-12T01:43:28,372 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-12T01:43:28,372 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:28,374 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-12T01:43:28,377 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-12T01:43:28,382 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:28,382 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:28,387 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:51152 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:43843:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51152 dst: /127.0.0.1:43843 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:28,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-12T01:43:28,395 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:28,398 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => ea8bffbd5aab17790fb7e6181e73636b, NAME => 'TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38 2024-12-12T01:43:28,404 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:28,404 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:28,408 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:52844 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:39973:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52844 dst: /127.0.0.1:39973 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:28,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-12T01:43:28,416 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:28,417 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:28,417 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing ea8bffbd5aab17790fb7e6181e73636b, disabling compactions & flushes 2024-12-12T01:43:28,417 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:28,417 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:28,417 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. after waiting 0 ms 2024-12-12T01:43:28,417 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:28,417 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:28,417 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for ea8bffbd5aab17790fb7e6181e73636b: Waiting for close lock at 1733967808417Disabling compacts and flushes for region at 1733967808417Disabling writes for close at 1733967808417Writing region close event to WAL at 1733967808417Closed at 1733967808417 2024-12-12T01:43:28,420 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-12T01:43:28,426 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733967808420"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733967808420"}]},"ts":"1733967808420"} 2024-12-12T01:43:28,431 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-12T01:43:28,433 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-12T01:43:28,436 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733967808433"}]},"ts":"1733967808433"} 2024-12-12T01:43:28,441 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-12T01:43:28,441 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {b85a53667e07=0} racks are {/default-rack=0} 2024-12-12T01:43:28,443 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-12T01:43:28,443 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-12T01:43:28,443 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-12T01:43:28,443 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-12T01:43:28,443 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-12T01:43:28,443 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-12T01:43:28,443 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-12T01:43:28,443 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-12T01:43:28,443 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-12T01:43:28,443 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-12T01:43:28,444 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ea8bffbd5aab17790fb7e6181e73636b, ASSIGN}] 2024-12-12T01:43:28,447 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ea8bffbd5aab17790fb7e6181e73636b, ASSIGN 2024-12-12T01:43:28,449 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ea8bffbd5aab17790fb7e6181e73636b, ASSIGN; state=OFFLINE, location=b85a53667e07,37251,1733967806152; forceNewPlan=false, retain=false 2024-12-12T01:43:28,493 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-12T01:43:28,602 INFO [b85a53667e07:36419 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-12T01:43:28,603 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=ea8bffbd5aab17790fb7e6181e73636b, regionState=OPENING, regionLocation=b85a53667e07,37251,1733967806152 2024-12-12T01:43:28,608 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ea8bffbd5aab17790fb7e6181e73636b, ASSIGN because future has completed 2024-12-12T01:43:28,610 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure ea8bffbd5aab17790fb7e6181e73636b, server=b85a53667e07,37251,1733967806152}] 2024-12-12T01:43:28,704 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-12T01:43:28,766 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-12T01:43:28,768 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56655, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-12T01:43:28,777 INFO [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:28,777 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => ea8bffbd5aab17790fb7e6181e73636b, NAME => 'TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b.', STARTKEY => '', ENDKEY => ''} 2024-12-12T01:43:28,777 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,777 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:28,778 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,778 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,780 INFO [StoreOpener-ea8bffbd5aab17790fb7e6181e73636b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,782 INFO [StoreOpener-ea8bffbd5aab17790fb7e6181e73636b-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region ea8bffbd5aab17790fb7e6181e73636b columnFamilyName cf 2024-12-12T01:43:28,782 DEBUG [StoreOpener-ea8bffbd5aab17790fb7e6181e73636b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:28,783 INFO [StoreOpener-ea8bffbd5aab17790fb7e6181e73636b-1 {}] regionserver.HStore(327): Store=ea8bffbd5aab17790fb7e6181e73636b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-12T01:43:28,783 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,785 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,785 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,786 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,786 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,788 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,793 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-12T01:43:28,795 INFO [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened ea8bffbd5aab17790fb7e6181e73636b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61524605, jitterRate=-0.08321194350719452}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-12T01:43:28,795 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:28,796 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for ea8bffbd5aab17790fb7e6181e73636b: Running coprocessor pre-open hook at 1733967808778Writing region info on filesystem at 1733967808778Initializing all the Stores at 1733967808779 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967808779Cleaning up temporary data from old regions at 1733967808786 (+7 ms)Running coprocessor post-open hooks at 1733967808795 (+9 ms)Region opened successfully at 1733967808796 (+1 ms) 2024-12-12T01:43:28,798 INFO [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b., pid=6, masterSystemTime=1733967808765 2024-12-12T01:43:28,803 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:28,803 INFO [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:28,804 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=ea8bffbd5aab17790fb7e6181e73636b, regionState=OPEN, openSeqNum=2, regionLocation=b85a53667e07,37251,1733967806152 2024-12-12T01:43:28,808 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure ea8bffbd5aab17790fb7e6181e73636b, server=b85a53667e07,37251,1733967806152 because future has completed 2024-12-12T01:43:28,814 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-12T01:43:28,814 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure ea8bffbd5aab17790fb7e6181e73636b, server=b85a53667e07,37251,1733967806152 in 200 msec 2024-12-12T01:43:28,817 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-12T01:43:28,818 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=ea8bffbd5aab17790fb7e6181e73636b, ASSIGN in 370 msec 2024-12-12T01:43:28,819 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-12T01:43:28,819 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733967808819"}]},"ts":"1733967808819"} 2024-12-12T01:43:28,823 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-12T01:43:28,825 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-12T01:43:28,829 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 462 msec 2024-12-12T01:43:29,013 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-12T01:43:29,014 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-12T01:43:29,014 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-12T01:43:29,016 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-12T01:43:29,023 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-12T01:43:29,024 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-12T01:43:29,024 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-12T01:43:29,033 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b., hostname=b85a53667e07,37251,1733967806152, seqNum=2] 2024-12-12T01:43:29,034 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-12T01:43:29,036 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34262, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-12T01:43:29,045 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-12T01:43:29,049 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-12T01:43:29,051 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-12T01:43:29,051 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-12T01:43:29,053 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-12T01:43:29,055 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-12T01:43:29,163 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-12T01:43:29,225 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37251 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-12T01:43:29,226 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:29,230 INFO [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing ea8bffbd5aab17790fb7e6181e73636b 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-12T01:43:29,277 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b/.tmp/cf/5a21374de08247e6b20eaa726da4f15c is 36, key is row/cf:cq/1733967809037/Put/seqid=0 2024-12-12T01:43:29,283 WARN [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:29,283 WARN [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:29,287 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1816902269_22 at /127.0.0.1:45486 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:41121:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45486 dst: /127.0.0.1:41121 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:29,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-12T01:43:29,292 WARN [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:29,292 INFO [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b/.tmp/cf/5a21374de08247e6b20eaa726da4f15c 2024-12-12T01:43:29,330 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b/.tmp/cf/5a21374de08247e6b20eaa726da4f15c as hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b/cf/5a21374de08247e6b20eaa726da4f15c 2024-12-12T01:43:29,341 INFO [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b/cf/5a21374de08247e6b20eaa726da4f15c, entries=1, sequenceid=5, filesize=4.7 K 2024-12-12T01:43:29,349 INFO [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for ea8bffbd5aab17790fb7e6181e73636b in 118ms, sequenceid=5, compaction requested=false 2024-12-12T01:43:29,350 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-12T01:43:29,352 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for ea8bffbd5aab17790fb7e6181e73636b: 2024-12-12T01:43:29,352 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:29,353 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-12T01:43:29,355 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-12T01:43:29,361 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-12T01:43:29,361 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 302 msec 2024-12-12T01:43:29,365 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 316 msec 2024-12-12T01:43:29,372 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36419 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-12T01:43:29,373 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-12T01:43:29,384 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-12T01:43:29,385 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-12T01:43:29,385 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:29,389 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:29,389 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:29,389 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-12T01:43:29,389 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-12T01:43:29,389 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1760953630, stopped=false 2024-12-12T01:43:29,390 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=b85a53667e07,36419,1733967805298 2024-12-12T01:43:29,444 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:29,444 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:29,444 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:29,444 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:29,444 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:29,444 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:29,444 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-12T01:43:29,444 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:29,444 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:29,445 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:29,445 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:29,445 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-12T01:43:29,446 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:29,446 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:29,446 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:29,446 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:29,446 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'b85a53667e07,37801,1733967805985' ***** 2024-12-12T01:43:29,447 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-12T01:43:29,447 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'b85a53667e07,46061,1733967806086' ***** 2024-12-12T01:43:29,447 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-12T01:43:29,447 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HeapMemoryManager(224): Stopping 2024-12-12T01:43:29,447 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'b85a53667e07,37251,1733967806152' ***** 2024-12-12T01:43:29,447 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-12T01:43:29,447 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HeapMemoryManager(224): Stopping 2024-12-12T01:43:29,448 INFO [RS:1;b85a53667e07:46061 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-12T01:43:29,448 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HeapMemoryManager(224): Stopping 2024-12-12T01:43:29,448 INFO [RS:0;b85a53667e07:37801 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-12T01:43:29,448 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-12T01:43:29,448 INFO [RS:2;b85a53667e07:37251 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-12T01:43:29,448 INFO [RS:1;b85a53667e07:46061 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-12T01:43:29,448 INFO [RS:2;b85a53667e07:37251 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-12T01:43:29,448 INFO [RS:0;b85a53667e07:37801 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-12T01:43:29,448 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(959): stopping server b85a53667e07,46061,1733967806086 2024-12-12T01:43:29,448 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(959): stopping server b85a53667e07,37801,1733967805985 2024-12-12T01:43:29,448 INFO [RS:1;b85a53667e07:46061 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-12T01:43:29,448 INFO [RS:0;b85a53667e07:37801 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-12T01:43:29,448 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(3091): Received CLOSE for ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:29,448 INFO [RS:1;b85a53667e07:46061 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;b85a53667e07:46061. 2024-12-12T01:43:29,448 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-12T01:43:29,448 INFO [RS:0;b85a53667e07:37801 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;b85a53667e07:37801. 2024-12-12T01:43:29,448 DEBUG [RS:1;b85a53667e07:46061 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:29,448 DEBUG [RS:1;b85a53667e07:46061 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:29,449 DEBUG [RS:0;b85a53667e07:37801 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:29,449 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-12T01:43:29,449 DEBUG [RS:0;b85a53667e07:37801 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:29,449 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(976): stopping server b85a53667e07,46061,1733967806086; all regions closed. 2024-12-12T01:43:29,449 INFO [RS:0;b85a53667e07:37801 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-12T01:43:29,449 INFO [RS:0;b85a53667e07:37801 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-12T01:43:29,449 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(959): stopping server b85a53667e07,37251,1733967806152 2024-12-12T01:43:29,449 INFO [RS:0;b85a53667e07:37801 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-12T01:43:29,449 INFO [RS:2;b85a53667e07:37251 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-12T01:43:29,449 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-12T01:43:29,449 INFO [RS:2;b85a53667e07:37251 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;b85a53667e07:37251. 2024-12-12T01:43:29,450 DEBUG [RS:2;b85a53667e07:37251 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:29,450 DEBUG [RS:2;b85a53667e07:37251 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:29,450 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-12T01:43:29,450 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-12T01:43:29,450 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-12T01:43:29,450 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-12T01:43:29,450 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1325): Online Regions={ea8bffbd5aab17790fb7e6181e73636b=TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b.} 2024-12-12T01:43:29,450 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing ea8bffbd5aab17790fb7e6181e73636b, disabling compactions & flushes 2024-12-12T01:43:29,450 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-12T01:43:29,450 INFO [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:29,450 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-12T01:43:29,450 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:29,450 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-12T01:43:29,450 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. after waiting 0 ms 2024-12-12T01:43:29,450 DEBUG [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1351): Waiting on ea8bffbd5aab17790fb7e6181e73636b 2024-12-12T01:43:29,450 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-12T01:43:29,450 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:29,451 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-12T01:43:29,451 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-12T01:43:29,451 INFO [regionserver/b85a53667e07:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-12T01:43:29,451 INFO [regionserver/b85a53667e07:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-12T01:43:29,452 INFO [regionserver/b85a53667e07:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-12T01:43:29,452 INFO [regionserver/b85a53667e07:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-12T01:43:29,453 INFO [regionserver/b85a53667e07:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:29,453 INFO [regionserver/b85a53667e07:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:29,453 INFO [regionserver/b85a53667e07:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:29,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_1073741827_1017 (size=93) 2024-12-12T01:43:29,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_1073741827_1017 (size=93) 2024-12-12T01:43:29,459 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_1073741827_1017 (size=93) 2024-12-12T01:43:29,465 DEBUG [RS:1;b85a53667e07:46061 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/oldWALs 2024-12-12T01:43:29,465 INFO [RS:1;b85a53667e07:46061 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL b85a53667e07%2C46061%2C1733967806086:(num 1733967807605) 2024-12-12T01:43:29,465 DEBUG [RS:1;b85a53667e07:46061 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:29,465 INFO [RS:1;b85a53667e07:46061 {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:29,465 INFO [RS:1;b85a53667e07:46061 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-12T01:43:29,465 INFO [RS:1;b85a53667e07:46061 {}] hbase.ChoreService(370): Chore service for: regionserver/b85a53667e07:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-12T01:43:29,466 INFO [RS:1;b85a53667e07:46061 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-12T01:43:29,466 INFO [RS:1;b85a53667e07:46061 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-12T01:43:29,466 INFO [RS:1;b85a53667e07:46061 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-12T01:43:29,466 INFO [regionserver/b85a53667e07:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-12T01:43:29,466 INFO [RS:1;b85a53667e07:46061 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-12T01:43:29,466 INFO [RS:1;b85a53667e07:46061 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:46061 2024-12-12T01:43:29,470 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/default/TestHBaseWalOnEC/ea8bffbd5aab17790fb7e6181e73636b/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-12T01:43:29,472 INFO [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:29,472 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for ea8bffbd5aab17790fb7e6181e73636b: Waiting for close lock at 1733967809449Running coprocessor pre-close hooks at 1733967809450 (+1 ms)Disabling compacts and flushes for region at 1733967809450Disabling writes for close at 1733967809450Writing region close event to WAL at 1733967809455 (+5 ms)Running coprocessor post-close hooks at 1733967809471 (+16 ms)Closed at 1733967809472 (+1 ms) 2024-12-12T01:43:29,473 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b. 2024-12-12T01:43:29,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/b85a53667e07,46061,1733967806086 2024-12-12T01:43:29,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-12T01:43:29,477 INFO [RS:1;b85a53667e07:46061 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-12T01:43:29,479 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/info/6f95b15ac0b6476589aef147099a6ad3 is 153, key is TestHBaseWalOnEC,,1733967808357.ea8bffbd5aab17790fb7e6181e73636b./info:regioninfo/1733967808804/Put/seqid=0 2024-12-12T01:43:29,482 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:29,482 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:29,482 INFO [regionserver/b85a53667e07:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-12T01:43:29,482 INFO [regionserver/b85a53667e07:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-12T01:43:29,486 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [b85a53667e07,46061,1733967806086] 2024-12-12T01:43:29,486 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1507619811_22 at /127.0.0.1:45500 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:41121:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45500 dst: /127.0.0.1:41121 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:29,490 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-12T01:43:29,493 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/b85a53667e07,46061,1733967806086 already deleted, retry=false 2024-12-12T01:43:29,493 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; b85a53667e07,46061,1733967806086 expired; onlineServers=2 2024-12-12T01:43:29,586 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:29,587 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46061-0x10017beb7220002, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:29,587 INFO [RS:1;b85a53667e07:46061 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-12T01:43:29,588 INFO [RS:1;b85a53667e07:46061 {}] regionserver.HRegionServer(1031): Exiting; stopping=b85a53667e07,46061,1733967806086; zookeeper connection closed. 2024-12-12T01:43:29,589 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5794307c {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5794307c 2024-12-12T01:43:29,651 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(976): stopping server b85a53667e07,37251,1733967806152; all regions closed. 2024-12-12T01:43:29,651 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-12T01:43:29,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_1073741826_1016 (size=1298) 2024-12-12T01:43:29,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_1073741826_1016 (size=1298) 2024-12-12T01:43:29,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_1073741826_1016 (size=1298) 2024-12-12T01:43:29,664 DEBUG [RS:2;b85a53667e07:37251 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/oldWALs 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL b85a53667e07%2C37251%2C1733967806152:(num 1733967807605) 2024-12-12T01:43:29,665 DEBUG [RS:2;b85a53667e07:37251 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] hbase.ChoreService(370): Chore service for: regionserver/b85a53667e07:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-12T01:43:29,665 INFO [regionserver/b85a53667e07:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-12T01:43:29,665 INFO [RS:2;b85a53667e07:37251 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:37251 2024-12-12T01:43:29,685 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/b85a53667e07,37251,1733967806152 2024-12-12T01:43:29,685 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-12T01:43:29,685 INFO [RS:2;b85a53667e07:37251 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-12T01:43:29,694 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [b85a53667e07,37251,1733967806152] 2024-12-12T01:43:29,702 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/b85a53667e07,37251,1733967806152 already deleted, retry=false 2024-12-12T01:43:29,702 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; b85a53667e07,37251,1733967806152 expired; onlineServers=1 2024-12-12T01:43:29,794 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:29,794 INFO [RS:2;b85a53667e07:37251 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-12T01:43:29,794 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37251-0x10017beb7220003, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:29,794 INFO [RS:2;b85a53667e07:37251 {}] regionserver.HRegionServer(1031): Exiting; stopping=b85a53667e07,37251,1733967806152; zookeeper connection closed. 2024-12-12T01:43:29,794 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7128094f {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7128094f 2024-12-12T01:43:29,851 DEBUG [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-12T01:43:29,893 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:29,894 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/info/6f95b15ac0b6476589aef147099a6ad3 2024-12-12T01:43:29,924 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/ns/73f5f92663374d6bbdefd0bb0378fb23 is 43, key is default/ns:d/1733967808193/Put/seqid=0 2024-12-12T01:43:29,927 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:29,927 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:29,932 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1507619811_22 at /127.0.0.1:51176 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:43843:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51176 dst: /127.0.0.1:43843 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:29,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-12T01:43:29,939 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:29,939 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/ns/73f5f92663374d6bbdefd0bb0378fb23 2024-12-12T01:43:29,964 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/table/eb41c927e5a6418f9c9a0092c3553d36 is 52, key is TestHBaseWalOnEC/table:state/1733967808819/Put/seqid=0 2024-12-12T01:43:29,966 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:29,966 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:29,969 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1507619811_22 at /127.0.0.1:52872 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:39973:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52872 dst: /127.0.0.1:39973 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:29,973 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-12T01:43:29,973 WARN [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:29,973 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/table/eb41c927e5a6418f9c9a0092c3553d36 2024-12-12T01:43:29,982 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/info/6f95b15ac0b6476589aef147099a6ad3 as hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/info/6f95b15ac0b6476589aef147099a6ad3 2024-12-12T01:43:29,991 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/info/6f95b15ac0b6476589aef147099a6ad3, entries=10, sequenceid=11, filesize=6.5 K 2024-12-12T01:43:29,992 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/ns/73f5f92663374d6bbdefd0bb0378fb23 as hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/ns/73f5f92663374d6bbdefd0bb0378fb23 2024-12-12T01:43:30,001 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/ns/73f5f92663374d6bbdefd0bb0378fb23, entries=2, sequenceid=11, filesize=5.0 K 2024-12-12T01:43:30,003 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/.tmp/table/eb41c927e5a6418f9c9a0092c3553d36 as hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/table/eb41c927e5a6418f9c9a0092c3553d36 2024-12-12T01:43:30,011 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/table/eb41c927e5a6418f9c9a0092c3553d36, entries=2, sequenceid=11, filesize=5.1 K 2024-12-12T01:43:30,012 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 562ms, sequenceid=11, compaction requested=false 2024-12-12T01:43:30,013 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-12T01:43:30,021 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-12T01:43:30,022 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-12T01:43:30,022 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-12T01:43:30,022 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733967809450Running coprocessor pre-close hooks at 1733967809450Disabling compacts and flushes for region at 1733967809450Disabling writes for close at 1733967809450Obtaining lock to block concurrent updates at 1733967809451 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1733967809451Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733967809451Flushing stores of hbase:meta,,1.1588230740 at 1733967809452 (+1 ms)Flushing 1588230740/info: creating writer at 1733967809452Flushing 1588230740/info: appending metadata at 1733967809475 (+23 ms)Flushing 1588230740/info: closing flushed file at 1733967809476 (+1 ms)Flushing 1588230740/ns: creating writer at 1733967809909 (+433 ms)Flushing 1588230740/ns: appending metadata at 1733967809923 (+14 ms)Flushing 1588230740/ns: closing flushed file at 1733967809923Flushing 1588230740/table: creating writer at 1733967809947 (+24 ms)Flushing 1588230740/table: appending metadata at 1733967809963 (+16 ms)Flushing 1588230740/table: closing flushed file at 1733967809963Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2a1e9992: reopening flushed file at 1733967809981 (+18 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5902a36b: reopening flushed file at 1733967809991 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5840ddb9: reopening flushed file at 1733967810001 (+10 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 562ms, sequenceid=11, compaction requested=false at 1733967810013 (+12 ms)Writing region close event to WAL at 1733967810014 (+1 ms)Running coprocessor post-close hooks at 1733967810022 (+8 ms)Closed at 1733967810022 2024-12-12T01:43:30,022 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-12T01:43:30,052 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(976): stopping server b85a53667e07,37801,1733967805985; all regions closed. 2024-12-12T01:43:30,056 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_1073741829_1019 (size=2751) 2024-12-12T01:43:30,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_1073741829_1019 (size=2751) 2024-12-12T01:43:30,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_1073741829_1019 (size=2751) 2024-12-12T01:43:30,062 DEBUG [RS:0;b85a53667e07:37801 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/oldWALs 2024-12-12T01:43:30,062 INFO [RS:0;b85a53667e07:37801 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL b85a53667e07%2C37801%2C1733967805985.meta:.meta(num 1733967808028) 2024-12-12T01:43:30,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_1073741828_1018 (size=93) 2024-12-12T01:43:30,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_1073741828_1018 (size=93) 2024-12-12T01:43:30,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_1073741828_1018 (size=93) 2024-12-12T01:43:30,069 DEBUG [RS:0;b85a53667e07:37801 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/oldWALs 2024-12-12T01:43:30,069 INFO [RS:0;b85a53667e07:37801 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL b85a53667e07%2C37801%2C1733967805985:(num 1733967807605) 2024-12-12T01:43:30,069 DEBUG [RS:0;b85a53667e07:37801 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:30,069 INFO [RS:0;b85a53667e07:37801 {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:30,069 INFO [RS:0;b85a53667e07:37801 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-12T01:43:30,069 INFO [RS:0;b85a53667e07:37801 {}] hbase.ChoreService(370): Chore service for: regionserver/b85a53667e07:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-12T01:43:30,070 INFO [RS:0;b85a53667e07:37801 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-12T01:43:30,070 INFO [regionserver/b85a53667e07:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-12T01:43:30,070 INFO [RS:0;b85a53667e07:37801 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:37801 2024-12-12T01:43:30,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/b85a53667e07,37801,1733967805985 2024-12-12T01:43:30,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-12T01:43:30,119 INFO [RS:0;b85a53667e07:37801 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-12T01:43:30,127 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [b85a53667e07,37801,1733967805985] 2024-12-12T01:43:30,135 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/b85a53667e07,37801,1733967805985 already deleted, retry=false 2024-12-12T01:43:30,135 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; b85a53667e07,37801,1733967805985 expired; onlineServers=0 2024-12-12T01:43:30,136 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'b85a53667e07,36419,1733967805298' ***** 2024-12-12T01:43:30,136 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-12T01:43:30,136 INFO [M:0;b85a53667e07:36419 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-12T01:43:30,137 INFO [M:0;b85a53667e07:36419 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-12T01:43:30,137 DEBUG [M:0;b85a53667e07:36419 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-12T01:43:30,137 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-12T01:43:30,137 DEBUG [M:0;b85a53667e07:36419 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-12T01:43:30,137 DEBUG [master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.large.0-1733967807181 {}] cleaner.HFileCleaner(306): Exit Thread[master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.large.0-1733967807181,5,FailOnTimeoutGroup] 2024-12-12T01:43:30,137 DEBUG [master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.small.0-1733967807182 {}] cleaner.HFileCleaner(306): Exit Thread[master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.small.0-1733967807182,5,FailOnTimeoutGroup] 2024-12-12T01:43:30,138 INFO [M:0;b85a53667e07:36419 {}] hbase.ChoreService(370): Chore service for: master/b85a53667e07:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-12T01:43:30,138 INFO [M:0;b85a53667e07:36419 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-12T01:43:30,138 DEBUG [M:0;b85a53667e07:36419 {}] master.HMaster(1795): Stopping service threads 2024-12-12T01:43:30,139 INFO [M:0;b85a53667e07:36419 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-12T01:43:30,139 INFO [M:0;b85a53667e07:36419 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-12T01:43:30,139 INFO [M:0;b85a53667e07:36419 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-12T01:43:30,140 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-12T01:43:30,143 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-12T01:43:30,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:30,144 DEBUG [M:0;b85a53667e07:36419 {}] zookeeper.ZKUtil(347): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-12T01:43:30,144 WARN [M:0;b85a53667e07:36419 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-12T01:43:30,145 INFO [M:0;b85a53667e07:36419 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/.lastflushedseqids 2024-12-12T01:43:30,157 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:30,158 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:30,160 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:45518 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:41121:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45518 dst: /127.0.0.1:41121 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:30,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-12T01:43:30,164 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:30,164 INFO [M:0;b85a53667e07:36419 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-12T01:43:30,165 INFO [M:0;b85a53667e07:36419 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-12T01:43:30,165 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-12T01:43:30,165 INFO [M:0;b85a53667e07:36419 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:30,165 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:30,165 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-12T01:43:30,165 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:30,165 INFO [M:0;b85a53667e07:36419 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.85 KB heapSize=34.13 KB 2024-12-12T01:43:30,182 DEBUG [M:0;b85a53667e07:36419 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/6502318b47b546b7bd4640b8dc539563 is 82, key is hbase:meta,,1/info:regioninfo/1733967808104/Put/seqid=0 2024-12-12T01:43:30,185 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:30,185 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:30,188 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:45530 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:41121:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45530 dst: /127.0.0.1:41121 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:30,191 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-12T01:43:30,192 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:30,192 INFO [M:0;b85a53667e07:36419 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/6502318b47b546b7bd4640b8dc539563 2024-12-12T01:43:30,213 DEBUG [M:0;b85a53667e07:36419 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/0b994d7991b44676ac3e709da6616c6b is 749, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733967808828/Put/seqid=0 2024-12-12T01:43:30,215 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:30,215 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:30,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775709_1013 (size=1321) 2024-12-12T01:43:30,220 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775708_1013 (size=1321) 2024-12-12T01:43:30,222 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:51194 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:43843:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51194 dst: /127.0.0.1:43843 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:30,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_-9223372036854775756_1006 (size=196) 2024-12-12T01:43:30,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775757_1006 (size=196) 2024-12-12T01:43:30,226 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775773_1004 (size=42) 2024-12-12T01:43:30,227 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:30,227 INFO [RS:0;b85a53667e07:37801 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-12T01:43:30,227 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37801-0x10017beb7220001, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:30,227 INFO [RS:0;b85a53667e07:37801 {}] regionserver.HRegionServer(1031): Exiting; stopping=b85a53667e07,37801,1733967805985; zookeeper connection closed. 2024-12-12T01:43:30,228 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775740_1008 (size=1189) 2024-12-12T01:43:30,228 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_-9223372036854775772_1004 (size=42) 2024-12-12T01:43:30,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_-9223372036854775741_1008 (size=1189) 2024-12-12T01:43:30,230 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5dde6ebf {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5dde6ebf 2024-12-12T01:43:30,230 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-12T01:43:30,231 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775552_1037 (size=6441) 2024-12-12T01:43:30,232 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:30,232 INFO [M:0;b85a53667e07:36419 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.17 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/0b994d7991b44676ac3e709da6616c6b 2024-12-12T01:43:30,254 DEBUG [M:0;b85a53667e07:36419 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/53ce14495c1f482696f95424ca29153b is 69, key is b85a53667e07,37251,1733967806152/rs:state/1733967807317/Put/seqid=0 2024-12-12T01:43:30,256 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:30,256 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-12T01:43:30,258 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1399083093_22 at /127.0.0.1:51242 [Receiving block BP-986496953-172.17.0.2-1733967801420:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:43843:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51242 dst: /127.0.0.1:43843 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-12T01:43:30,262 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-12T01:43:30,263 WARN [M:0;b85a53667e07:36419 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-12T01:43:30,263 INFO [M:0;b85a53667e07:36419 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/53ce14495c1f482696f95424ca29153b 2024-12-12T01:43:30,271 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/6502318b47b546b7bd4640b8dc539563 as hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/6502318b47b546b7bd4640b8dc539563 2024-12-12T01:43:30,278 INFO [M:0;b85a53667e07:36419 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/6502318b47b546b7bd4640b8dc539563, entries=8, sequenceid=72, filesize=5.5 K 2024-12-12T01:43:30,279 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/0b994d7991b44676ac3e709da6616c6b as hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/0b994d7991b44676ac3e709da6616c6b 2024-12-12T01:43:30,287 INFO [M:0;b85a53667e07:36419 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/0b994d7991b44676ac3e709da6616c6b, entries=8, sequenceid=72, filesize=6.3 K 2024-12-12T01:43:30,289 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/53ce14495c1f482696f95424ca29153b as hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/53ce14495c1f482696f95424ca29153b 2024-12-12T01:43:30,296 INFO [M:0;b85a53667e07:36419 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/53ce14495c1f482696f95424ca29153b, entries=3, sequenceid=72, filesize=5.2 K 2024-12-12T01:43:30,297 INFO [M:0;b85a53667e07:36419 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 132ms, sequenceid=72, compaction requested=false 2024-12-12T01:43:30,298 INFO [M:0;b85a53667e07:36419 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:30,299 DEBUG [M:0;b85a53667e07:36419 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733967810165Disabling compacts and flushes for region at 1733967810165Disabling writes for close at 1733967810165Obtaining lock to block concurrent updates at 1733967810165Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733967810165Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27492, getHeapSize=34888, getOffHeapSize=0, getCellsCount=85 at 1733967810166 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733967810166Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733967810167 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733967810181 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733967810181Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733967810198 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733967810212 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733967810212Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733967810239 (+27 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733967810254 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733967810254Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5874c5ed: reopening flushed file at 1733967810270 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5cbb3a2e: reopening flushed file at 1733967810278 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3146ecc8: reopening flushed file at 1733967810287 (+9 ms)Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 132ms, sequenceid=72, compaction requested=false at 1733967810297 (+10 ms)Writing region close event to WAL at 1733967810298 (+1 ms)Closed at 1733967810298 2024-12-12T01:43:30,302 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41121 is added to blk_1073741825_1011 (size=32695) 2024-12-12T01:43:30,302 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43843 is added to blk_1073741825_1011 (size=32695) 2024-12-12T01:43:30,302 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39973 is added to blk_1073741825_1011 (size=32695) 2024-12-12T01:43:30,303 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-12T01:43:30,303 INFO [M:0;b85a53667e07:36419 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-12T01:43:30,303 INFO [M:0;b85a53667e07:36419 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36419 2024-12-12T01:43:30,303 INFO [M:0;b85a53667e07:36419 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-12T01:43:30,427 INFO [M:0;b85a53667e07:36419 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-12T01:43:30,427 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:30,427 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36419-0x10017beb7220000, quorum=127.0.0.1:60989, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:30,435 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@793c26cb{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:30,439 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3e0ab7c3{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-12T01:43:30,439 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-12T01:43:30,439 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@36ef30b2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-12T01:43:30,439 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@f4b982a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir/,STOPPED} 2024-12-12T01:43:30,442 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-12T01:43:30,442 WARN [BP-986496953-172.17.0.2-1733967801420 heartbeating to localhost/127.0.0.1:40607 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-12T01:43:30,442 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-12T01:43:30,442 WARN [BP-986496953-172.17.0.2-1733967801420 heartbeating to localhost/127.0.0.1:40607 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-986496953-172.17.0.2-1733967801420 (Datanode Uuid 5ec7a961-dbb5-4343-9349-880ff4c9a610) service to localhost/127.0.0.1:40607 2024-12-12T01:43:30,443 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data5/current/BP-986496953-172.17.0.2-1733967801420 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:30,443 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data6/current/BP-986496953-172.17.0.2-1733967801420 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:30,444 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-12T01:43:30,445 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@43fe7c8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:30,446 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@6fecef94{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-12T01:43:30,446 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-12T01:43:30,446 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1e752fdd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-12T01:43:30,446 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3f04a93b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir/,STOPPED} 2024-12-12T01:43:30,447 WARN [BP-986496953-172.17.0.2-1733967801420 heartbeating to localhost/127.0.0.1:40607 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-12T01:43:30,447 WARN [BP-986496953-172.17.0.2-1733967801420 heartbeating to localhost/127.0.0.1:40607 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-986496953-172.17.0.2-1733967801420 (Datanode Uuid 132dc7f6-a5d8-460e-b826-15809fcc7046) service to localhost/127.0.0.1:40607 2024-12-12T01:43:30,448 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data3/current/BP-986496953-172.17.0.2-1733967801420 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:30,448 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data4/current/BP-986496953-172.17.0.2-1733967801420 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:30,448 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-12T01:43:30,448 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-12T01:43:30,448 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-12T01:43:30,450 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@103787ab{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:30,450 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@23a1a692{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-12T01:43:30,450 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-12T01:43:30,450 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@79f7513c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-12T01:43:30,451 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2d15633c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir/,STOPPED} 2024-12-12T01:43:30,452 WARN [BP-986496953-172.17.0.2-1733967801420 heartbeating to localhost/127.0.0.1:40607 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-12T01:43:30,452 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-12T01:43:30,452 WARN [BP-986496953-172.17.0.2-1733967801420 heartbeating to localhost/127.0.0.1:40607 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-986496953-172.17.0.2-1733967801420 (Datanode Uuid 03106b5e-0306-482f-b556-190a78e07bb5) service to localhost/127.0.0.1:40607 2024-12-12T01:43:30,452 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-12T01:43:30,452 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data1/current/BP-986496953-172.17.0.2-1733967801420 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:30,452 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/cluster_8fd1b054-1e6a-f460-3959-7c4ab824a647/data/data2/current/BP-986496953-172.17.0.2-1733967801420 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:30,453 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-12T01:43:30,460 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@63603efd{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-12T01:43:30,460 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3f607d40{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-12T01:43:30,460 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-12T01:43:30,460 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@e67f021{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-12T01:43:30,461 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@77f7f078{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir/,STOPPED} 2024-12-12T01:43:30,469 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-12T01:43:30,494 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-12T01:43:30,500 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=87 (was 157), OpenFileDescriptor=439 (was 393) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=329 (was 349), ProcessCount=11 (was 11), AvailableMemoryMB=9447 (was 9713) 2024-12-12T01:43:30,505 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=87, OpenFileDescriptor=439, MaxFileDescriptor=1048576, SystemLoadAverage=329, ProcessCount=11, AvailableMemoryMB=9447 2024-12-12T01:43:30,505 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-12T01:43:30,505 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.log.dir so I do NOT create it in target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f 2024-12-12T01:43:30,505 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/86c1d766-b389-5ecc-3db8-e9e29593ccb7/hadoop.tmp.dir so I do NOT create it in target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f 2024-12-12T01:43:30,505 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c, deleteOnExit=true 2024-12-12T01:43:30,505 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/test.cache.data in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.tmp.dir in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-12T01:43:30,506 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-12T01:43:30,506 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/nfs.dump.dir in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/java.io.tmpdir in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-12T01:43:30,507 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-12T01:43:30,773 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:30,779 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-12T01:43:30,787 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-12T01:43:30,787 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-12T01:43:30,787 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-12T01:43:30,788 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:30,789 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@42b66dc7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir/,AVAILABLE} 2024-12-12T01:43:30,789 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7940a36d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-12T01:43:30,879 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@391ced48{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/java.io.tmpdir/jetty-localhost-45325-hadoop-hdfs-3_4_1-tests_jar-_-any-3436570571981684932/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-12T01:43:30,880 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@38b34206{HTTP/1.1, (http/1.1)}{localhost:45325} 2024-12-12T01:43:30,880 INFO [Time-limited test {}] server.Server(415): Started @11080ms 2024-12-12T01:43:31,100 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:31,103 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-12T01:43:31,104 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-12T01:43:31,104 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-12T01:43:31,104 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-12T01:43:31,105 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3dec3425{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir/,AVAILABLE} 2024-12-12T01:43:31,105 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@56d7456a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-12T01:43:31,196 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@24fac726{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/java.io.tmpdir/jetty-localhost-40203-hadoop-hdfs-3_4_1-tests_jar-_-any-15344692188448555268/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:31,197 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11c2a472{HTTP/1.1, (http/1.1)}{localhost:40203} 2024-12-12T01:43:31,197 INFO [Time-limited test {}] server.Server(415): Started @11397ms 2024-12-12T01:43:31,198 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-12T01:43:31,228 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:31,230 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-12T01:43:31,231 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-12T01:43:31,231 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-12T01:43:31,231 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-12T01:43:31,231 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5a1d6627{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir/,AVAILABLE} 2024-12-12T01:43:31,232 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@24b4498d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-12T01:43:31,321 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@31df22e4{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/java.io.tmpdir/jetty-localhost-39773-hadoop-hdfs-3_4_1-tests_jar-_-any-16133622167621349063/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:31,322 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1c756a11{HTTP/1.1, (http/1.1)}{localhost:39773} 2024-12-12T01:43:31,322 INFO [Time-limited test {}] server.Server(415): Started @11522ms 2024-12-12T01:43:31,323 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-12T01:43:31,350 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-12T01:43:31,352 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-12T01:43:31,353 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-12T01:43:31,353 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-12T01:43:31,353 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-12T01:43:31,353 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@74dea45e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir/,AVAILABLE} 2024-12-12T01:43:31,354 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@40bb076c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-12T01:43:31,442 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@53d1de4b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/java.io.tmpdir/jetty-localhost-33227-hadoop-hdfs-3_4_1-tests_jar-_-any-6439458588394863816/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:31,442 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@584937f8{HTTP/1.1, (http/1.1)}{localhost:33227} 2024-12-12T01:43:31,442 INFO [Time-limited test {}] server.Server(415): Started @11642ms 2024-12-12T01:43:31,443 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-12T01:43:31,969 WARN [Thread-560 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data1/current/BP-2131542035-172.17.0.2-1733967810531/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:31,970 WARN [Thread-561 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data2/current/BP-2131542035-172.17.0.2-1733967810531/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:31,987 WARN [Thread-502 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-12T01:43:31,990 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc6124c78cee2ceca with lease ID 0xef79ccf4d1e456b4: Processing first storage report for DS-e6960683-d5a2-4e5e-a7f8-7bd49f1fff6b from datanode DatanodeRegistration(127.0.0.1:38519, datanodeUuid=f4693875-b11f-4404-9a8e-9e903552c827, infoPort=41387, infoSecurePort=0, ipcPort=46755, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531) 2024-12-12T01:43:31,990 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc6124c78cee2ceca with lease ID 0xef79ccf4d1e456b4: from storage DS-e6960683-d5a2-4e5e-a7f8-7bd49f1fff6b node DatanodeRegistration(127.0.0.1:38519, datanodeUuid=f4693875-b11f-4404-9a8e-9e903552c827, infoPort=41387, infoSecurePort=0, ipcPort=46755, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:31,990 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc6124c78cee2ceca with lease ID 0xef79ccf4d1e456b4: Processing first storage report for DS-f7cfbccf-f00a-4876-9fdd-9bb9b51a27f6 from datanode DatanodeRegistration(127.0.0.1:38519, datanodeUuid=f4693875-b11f-4404-9a8e-9e903552c827, infoPort=41387, infoSecurePort=0, ipcPort=46755, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531) 2024-12-12T01:43:31,990 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc6124c78cee2ceca with lease ID 0xef79ccf4d1e456b4: from storage DS-f7cfbccf-f00a-4876-9fdd-9bb9b51a27f6 node DatanodeRegistration(127.0.0.1:38519, datanodeUuid=f4693875-b11f-4404-9a8e-9e903552c827, infoPort=41387, infoSecurePort=0, ipcPort=46755, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:32,273 WARN [Thread-573 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data3/current/BP-2131542035-172.17.0.2-1733967810531/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:32,273 WARN [Thread-574 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data4/current/BP-2131542035-172.17.0.2-1733967810531/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:32,289 WARN [Thread-525 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-12T01:43:32,292 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd5a7353bf3bbdcca with lease ID 0xef79ccf4d1e456b5: Processing first storage report for DS-43ddbed3-216b-46c4-9357-a129d6e92e1d from datanode DatanodeRegistration(127.0.0.1:45293, datanodeUuid=4caee6f0-92fb-489e-a0bf-c2a9ca308efd, infoPort=38777, infoSecurePort=0, ipcPort=39471, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531) 2024-12-12T01:43:32,292 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd5a7353bf3bbdcca with lease ID 0xef79ccf4d1e456b5: from storage DS-43ddbed3-216b-46c4-9357-a129d6e92e1d node DatanodeRegistration(127.0.0.1:45293, datanodeUuid=4caee6f0-92fb-489e-a0bf-c2a9ca308efd, infoPort=38777, infoSecurePort=0, ipcPort=39471, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:32,292 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd5a7353bf3bbdcca with lease ID 0xef79ccf4d1e456b5: Processing first storage report for DS-f7c9376c-1f6e-4514-994d-99057d988a0a from datanode DatanodeRegistration(127.0.0.1:45293, datanodeUuid=4caee6f0-92fb-489e-a0bf-c2a9ca308efd, infoPort=38777, infoSecurePort=0, ipcPort=39471, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531) 2024-12-12T01:43:32,292 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd5a7353bf3bbdcca with lease ID 0xef79ccf4d1e456b5: from storage DS-f7c9376c-1f6e-4514-994d-99057d988a0a node DatanodeRegistration(127.0.0.1:45293, datanodeUuid=4caee6f0-92fb-489e-a0bf-c2a9ca308efd, infoPort=38777, infoSecurePort=0, ipcPort=39471, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:32,346 WARN [Thread-584 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data5/current/BP-2131542035-172.17.0.2-1733967810531/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:32,346 WARN [Thread-585 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data6/current/BP-2131542035-172.17.0.2-1733967810531/current, will proceed with Du for space computation calculation, 2024-12-12T01:43:32,364 WARN [Thread-547 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-12T01:43:32,366 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x72c16194c7400d45 with lease ID 0xef79ccf4d1e456b6: Processing first storage report for DS-4c6a85ba-c04c-4eef-a52b-1605a21cadef from datanode DatanodeRegistration(127.0.0.1:39739, datanodeUuid=fcb0eafa-82f6-4c36-9420-5c0bb0b50f5a, infoPort=39021, infoSecurePort=0, ipcPort=35253, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531) 2024-12-12T01:43:32,366 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x72c16194c7400d45 with lease ID 0xef79ccf4d1e456b6: from storage DS-4c6a85ba-c04c-4eef-a52b-1605a21cadef node DatanodeRegistration(127.0.0.1:39739, datanodeUuid=fcb0eafa-82f6-4c36-9420-5c0bb0b50f5a, infoPort=39021, infoSecurePort=0, ipcPort=35253, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-12T01:43:32,367 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x72c16194c7400d45 with lease ID 0xef79ccf4d1e456b6: Processing first storage report for DS-728dcd87-f709-4ea1-ba94-7cfdfd00d2e7 from datanode DatanodeRegistration(127.0.0.1:39739, datanodeUuid=fcb0eafa-82f6-4c36-9420-5c0bb0b50f5a, infoPort=39021, infoSecurePort=0, ipcPort=35253, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531) 2024-12-12T01:43:32,367 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x72c16194c7400d45 with lease ID 0xef79ccf4d1e456b6: from storage DS-728dcd87-f709-4ea1-ba94-7cfdfd00d2e7 node DatanodeRegistration(127.0.0.1:39739, datanodeUuid=fcb0eafa-82f6-4c36-9420-5c0bb0b50f5a, infoPort=39021, infoSecurePort=0, ipcPort=35253, storageInfo=lv=-57;cid=testClusterID;nsid=1012130605;c=1733967810531), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-12T01:43:32,378 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f 2024-12-12T01:43:32,381 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/zookeeper_0, clientPort=63173, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-12T01:43:32,382 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=63173 2024-12-12T01:43:32,382 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,384 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741825_1001 (size=7) 2024-12-12T01:43:32,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741825_1001 (size=7) 2024-12-12T01:43:32,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741825_1001 (size=7) 2024-12-12T01:43:32,397 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca with version=8 2024-12-12T01:43:32,397 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:40607/user/jenkins/test-data/e5d416ea-6439-265f-73e7-3cc3fba09b38/hbase-staging 2024-12-12T01:43:32,398 INFO [Time-limited test {}] client.ConnectionUtils(128): master/b85a53667e07:0 server-side Connection retries=45 2024-12-12T01:43:32,398 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,399 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,399 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-12T01:43:32,399 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,399 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-12T01:43:32,399 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-12T01:43:32,399 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-12T01:43:32,400 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41417 2024-12-12T01:43:32,401 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:41417 connecting to ZooKeeper ensemble=127.0.0.1:63173 2024-12-12T01:43:32,451 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:414170x0, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-12T01:43:32,451 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:41417-0x10017bed5b90000 connected 2024-12-12T01:43:32,520 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,524 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,527 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:32,527 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca, hbase.cluster.distributed=false 2024-12-12T01:43:32,530 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-12T01:43:32,531 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41417 2024-12-12T01:43:32,531 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41417 2024-12-12T01:43:32,532 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41417 2024-12-12T01:43:32,532 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41417 2024-12-12T01:43:32,533 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41417 2024-12-12T01:43:32,550 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/b85a53667e07:0 server-side Connection retries=45 2024-12-12T01:43:32,550 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,550 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,550 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-12T01:43:32,550 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,550 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-12T01:43:32,550 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-12T01:43:32,550 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-12T01:43:32,551 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41049 2024-12-12T01:43:32,552 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:41049 connecting to ZooKeeper ensemble=127.0.0.1:63173 2024-12-12T01:43:32,553 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,554 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,567 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:410490x0, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-12T01:43:32,568 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:41049-0x10017bed5b90001 connected 2024-12-12T01:43:32,568 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:32,568 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-12T01:43:32,569 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-12T01:43:32,569 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-12T01:43:32,571 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-12T01:43:32,571 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41049 2024-12-12T01:43:32,572 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41049 2024-12-12T01:43:32,572 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41049 2024-12-12T01:43:32,573 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41049 2024-12-12T01:43:32,573 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41049 2024-12-12T01:43:32,588 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/b85a53667e07:0 server-side Connection retries=45 2024-12-12T01:43:32,588 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,588 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,588 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-12T01:43:32,588 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,588 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-12T01:43:32,588 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-12T01:43:32,589 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-12T01:43:32,589 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:43203 2024-12-12T01:43:32,590 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:43203 connecting to ZooKeeper ensemble=127.0.0.1:63173 2024-12-12T01:43:32,591 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,592 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,601 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:432030x0, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-12T01:43:32,602 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:43203-0x10017bed5b90002 connected 2024-12-12T01:43:32,602 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:32,602 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-12T01:43:32,603 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-12T01:43:32,603 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-12T01:43:32,605 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-12T01:43:32,605 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43203 2024-12-12T01:43:32,605 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43203 2024-12-12T01:43:32,605 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43203 2024-12-12T01:43:32,606 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43203 2024-12-12T01:43:32,606 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43203 2024-12-12T01:43:32,624 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/b85a53667e07:0 server-side Connection retries=45 2024-12-12T01:43:32,624 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,624 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,624 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-12T01:43:32,624 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-12T01:43:32,624 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-12T01:43:32,624 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-12T01:43:32,624 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-12T01:43:32,625 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41011 2024-12-12T01:43:32,626 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:41011 connecting to ZooKeeper ensemble=127.0.0.1:63173 2024-12-12T01:43:32,626 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,628 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,635 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:410110x0, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-12T01:43:32,635 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:41011-0x10017bed5b90003 connected 2024-12-12T01:43:32,635 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:32,636 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-12T01:43:32,636 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-12T01:43:32,637 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-12T01:43:32,638 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-12T01:43:32,639 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41011 2024-12-12T01:43:32,639 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41011 2024-12-12T01:43:32,640 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41011 2024-12-12T01:43:32,641 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41011 2024-12-12T01:43:32,641 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41011 2024-12-12T01:43:32,652 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;b85a53667e07:41417 2024-12-12T01:43:32,653 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/b85a53667e07,41417,1733967812398 2024-12-12T01:43:32,660 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:32,660 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:32,660 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:32,660 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:32,660 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/b85a53667e07,41417,1733967812398 2024-12-12T01:43:32,668 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-12T01:43:32,668 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-12T01:43:32,668 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-12T01:43:32,668 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,668 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,668 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,668 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,669 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-12T01:43:32,670 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/b85a53667e07,41417,1733967812398 from backup master directory 2024-12-12T01:43:32,676 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/b85a53667e07,41417,1733967812398 2024-12-12T01:43:32,676 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:32,676 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:32,677 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:32,677 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-12T01:43:32,677 WARN [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-12T01:43:32,677 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=b85a53667e07,41417,1733967812398 2024-12-12T01:43:32,686 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/hbase.id] with ID: 45f7c04c-dfa0-431f-9efc-18ac0abe519f 2024-12-12T01:43:32,686 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/.tmp/hbase.id 2024-12-12T01:43:32,700 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741826_1002 (size=42) 2024-12-12T01:43:32,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741826_1002 (size=42) 2024-12-12T01:43:32,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741826_1002 (size=42) 2024-12-12T01:43:32,702 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/.tmp/hbase.id]:[hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/hbase.id] 2024-12-12T01:43:32,717 INFO [master/b85a53667e07:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-12T01:43:32,717 INFO [master/b85a53667e07:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-12T01:43:32,718 INFO [master/b85a53667e07:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 1ms. 2024-12-12T01:43:32,726 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,726 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,726 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,726 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,736 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741827_1003 (size=196) 2024-12-12T01:43:32,736 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741827_1003 (size=196) 2024-12-12T01:43:32,736 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741827_1003 (size=196) 2024-12-12T01:43:32,739 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-12T01:43:32,739 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-12T01:43:32,740 INFO [master/b85a53667e07:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-12T01:43:32,750 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741828_1004 (size=1189) 2024-12-12T01:43:32,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741828_1004 (size=1189) 2024-12-12T01:43:32,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741828_1004 (size=1189) 2024-12-12T01:43:32,752 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store 2024-12-12T01:43:32,761 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741829_1005 (size=34) 2024-12-12T01:43:32,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741829_1005 (size=34) 2024-12-12T01:43:32,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741829_1005 (size=34) 2024-12-12T01:43:32,763 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:32,763 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-12T01:43:32,763 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:32,763 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:32,763 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-12T01:43:32,763 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:32,763 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:32,763 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733967812763Disabling compacts and flushes for region at 1733967812763Disabling writes for close at 1733967812763Writing region close event to WAL at 1733967812763Closed at 1733967812763 2024-12-12T01:43:32,764 WARN [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/.initializing 2024-12-12T01:43:32,764 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/WALs/b85a53667e07,41417,1733967812398 2024-12-12T01:43:32,768 INFO [master/b85a53667e07:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C41417%2C1733967812398, suffix=, logDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/WALs/b85a53667e07,41417,1733967812398, archiveDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/oldWALs, maxLogs=10 2024-12-12T01:43:32,769 INFO [master/b85a53667e07:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor b85a53667e07%2C41417%2C1733967812398.1733967812769 2024-12-12T01:43:32,779 INFO [master/b85a53667e07:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/WALs/b85a53667e07,41417,1733967812398/b85a53667e07%2C41417%2C1733967812398.1733967812769 2024-12-12T01:43:32,781 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41387:41387),(127.0.0.1/127.0.0.1:38777:38777),(127.0.0.1/127.0.0.1:39021:39021)] 2024-12-12T01:43:32,782 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-12T01:43:32,782 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:32,782 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,782 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,784 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,786 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-12T01:43:32,786 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:32,787 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:32,787 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,789 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-12T01:43:32,789 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:32,789 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-12T01:43:32,789 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,792 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-12T01:43:32,792 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:32,793 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-12T01:43:32,793 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,795 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-12T01:43:32,795 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:32,796 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-12T01:43:32,796 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,796 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,797 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,798 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,798 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,799 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-12T01:43:32,800 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-12T01:43:32,802 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-12T01:43:32,803 INFO [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65933745, jitterRate=-0.017510637640953064}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-12T01:43:32,803 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733967812783Initializing all the Stores at 1733967812784 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967812784Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967812784Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967812784Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967812784Cleaning up temporary data from old regions at 1733967812798 (+14 ms)Region opened successfully at 1733967812803 (+5 ms) 2024-12-12T01:43:32,804 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-12T01:43:32,807 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@605a79fa, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=b85a53667e07/172.17.0.2:0 2024-12-12T01:43:32,808 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-12T01:43:32,808 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-12T01:43:32,808 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-12T01:43:32,809 INFO [master/b85a53667e07:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-12T01:43:32,809 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-12T01:43:32,810 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-12T01:43:32,810 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-12T01:43:32,812 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-12T01:43:32,813 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-12T01:43:32,842 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-12T01:43:32,843 INFO [master/b85a53667e07:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-12T01:43:32,844 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-12T01:43:32,851 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-12T01:43:32,852 INFO [master/b85a53667e07:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-12T01:43:32,853 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-12T01:43:32,860 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-12T01:43:32,861 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-12T01:43:32,868 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-12T01:43:32,870 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-12T01:43:32,876 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-12T01:43:32,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:32,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:32,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:32,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:32,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,886 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=b85a53667e07,41417,1733967812398, sessionid=0x10017bed5b90000, setting cluster-up flag (Was=false) 2024-12-12T01:43:32,901 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,901 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,901 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,901 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,926 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-12T01:43:32,928 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=b85a53667e07,41417,1733967812398 2024-12-12T01:43:32,943 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,943 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,943 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,943 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:32,968 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-12T01:43:32,971 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=b85a53667e07,41417,1733967812398 2024-12-12T01:43:32,973 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-12T01:43:32,975 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-12T01:43:32,976 INFO [master/b85a53667e07:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-12T01:43:32,976 INFO [master/b85a53667e07:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-12T01:43:32,976 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: b85a53667e07,41417,1733967812398 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-12T01:43:32,978 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/b85a53667e07:0, corePoolSize=5, maxPoolSize=5 2024-12-12T01:43:32,978 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/b85a53667e07:0, corePoolSize=5, maxPoolSize=5 2024-12-12T01:43:32,978 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/b85a53667e07:0, corePoolSize=5, maxPoolSize=5 2024-12-12T01:43:32,978 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/b85a53667e07:0, corePoolSize=5, maxPoolSize=5 2024-12-12T01:43:32,978 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/b85a53667e07:0, corePoolSize=10, maxPoolSize=10 2024-12-12T01:43:32,978 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:32,978 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/b85a53667e07:0, corePoolSize=2, maxPoolSize=2 2024-12-12T01:43:32,978 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:32,979 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733967842979 2024-12-12T01:43:32,979 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-12T01:43:32,980 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-12T01:43:32,980 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-12T01:43:32,980 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-12T01:43:32,980 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-12T01:43:32,980 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-12T01:43:32,980 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:32,980 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-12T01:43:32,981 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-12T01:43:32,981 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-12T01:43:32,981 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-12T01:43:32,981 INFO [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-12T01:43:32,981 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-12T01:43:32,981 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-12T01:43:32,981 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.large.0-1733967812981,5,FailOnTimeoutGroup] 2024-12-12T01:43:32,981 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.small.0-1733967812981,5,FailOnTimeoutGroup] 2024-12-12T01:43:32,981 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:32,981 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-12T01:43:32,981 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:32,981 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:32,982 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:32,982 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-12T01:43:32,993 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741831_1007 (size=1321) 2024-12-12T01:43:32,993 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741831_1007 (size=1321) 2024-12-12T01:43:32,993 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741831_1007 (size=1321) 2024-12-12T01:43:32,994 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-12T01:43:32,994 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca 2024-12-12T01:43:33,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741832_1008 (size=32) 2024-12-12T01:43:33,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741832_1008 (size=32) 2024-12-12T01:43:33,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741832_1008 (size=32) 2024-12-12T01:43:33,005 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:33,007 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-12T01:43:33,008 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-12T01:43:33,008 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,009 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:33,009 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-12T01:43:33,010 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-12T01:43:33,010 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,011 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:33,011 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-12T01:43:33,012 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-12T01:43:33,012 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,013 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:33,013 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-12T01:43:33,014 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-12T01:43:33,014 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,015 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:33,015 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-12T01:43:33,016 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740 2024-12-12T01:43:33,016 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740 2024-12-12T01:43:33,018 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-12T01:43:33,018 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-12T01:43:33,019 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-12T01:43:33,020 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-12T01:43:33,023 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-12T01:43:33,024 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70442442, jitterRate=0.04967418313026428}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-12T01:43:33,025 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733967813005Initializing all the Stores at 1733967813006 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967813006Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967813007 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967813007Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967813007Cleaning up temporary data from old regions at 1733967813018 (+11 ms)Region opened successfully at 1733967813025 (+7 ms) 2024-12-12T01:43:33,025 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-12T01:43:33,025 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-12T01:43:33,025 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-12T01:43:33,025 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-12T01:43:33,025 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-12T01:43:33,026 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-12T01:43:33,026 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733967813025Disabling compacts and flushes for region at 1733967813025Disabling writes for close at 1733967813025Writing region close event to WAL at 1733967813025Closed at 1733967813025 2024-12-12T01:43:33,028 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-12T01:43:33,028 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-12T01:43:33,028 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-12T01:43:33,030 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-12T01:43:33,032 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-12T01:43:33,043 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(746): ClusterId : 45f7c04c-dfa0-431f-9efc-18ac0abe519f 2024-12-12T01:43:33,043 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(746): ClusterId : 45f7c04c-dfa0-431f-9efc-18ac0abe519f 2024-12-12T01:43:33,043 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(746): ClusterId : 45f7c04c-dfa0-431f-9efc-18ac0abe519f 2024-12-12T01:43:33,043 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-12T01:43:33,043 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-12T01:43:33,043 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-12T01:43:33,069 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-12T01:43:33,069 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-12T01:43:33,069 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-12T01:43:33,069 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-12T01:43:33,069 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-12T01:43:33,069 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-12T01:43:33,086 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-12T01:43:33,086 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-12T01:43:33,086 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-12T01:43:33,087 DEBUG [RS:1;b85a53667e07:43203 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@47c37a6f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=b85a53667e07/172.17.0.2:0 2024-12-12T01:43:33,087 DEBUG [RS:2;b85a53667e07:41011 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2011674c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=b85a53667e07/172.17.0.2:0 2024-12-12T01:43:33,087 DEBUG [RS:0;b85a53667e07:41049 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@516fc872, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=b85a53667e07/172.17.0.2:0 2024-12-12T01:43:33,101 DEBUG [RS:0;b85a53667e07:41049 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;b85a53667e07:41049 2024-12-12T01:43:33,101 INFO [RS:0;b85a53667e07:41049 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-12T01:43:33,101 INFO [RS:0;b85a53667e07:41049 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-12T01:43:33,101 DEBUG [RS:1;b85a53667e07:43203 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;b85a53667e07:43203 2024-12-12T01:43:33,101 DEBUG [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-12T01:43:33,101 INFO [RS:1;b85a53667e07:43203 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-12T01:43:33,101 INFO [RS:1;b85a53667e07:43203 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-12T01:43:33,101 DEBUG [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-12T01:43:33,102 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,41417,1733967812398 with port=41049, startcode=1733967812549 2024-12-12T01:43:33,102 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,41417,1733967812398 with port=43203, startcode=1733967812588 2024-12-12T01:43:33,102 DEBUG [RS:0;b85a53667e07:41049 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-12T01:43:33,102 DEBUG [RS:1;b85a53667e07:43203 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-12T01:43:33,105 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59637, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-12T01:43:33,105 DEBUG [RS:2;b85a53667e07:41011 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;b85a53667e07:41011 2024-12-12T01:43:33,105 INFO [RS:2;b85a53667e07:41011 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-12T01:43:33,105 INFO [RS:2;b85a53667e07:41011 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-12T01:43:33,105 DEBUG [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-12T01:43:33,105 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41417 {}] master.ServerManager(363): Checking decommissioned status of RegionServer b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,105 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41417 {}] master.ServerManager(517): Registering regionserver=b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,106 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(2659): reportForDuty to master=b85a53667e07,41417,1733967812398 with port=41011, startcode=1733967812624 2024-12-12T01:43:33,106 DEBUG [RS:2;b85a53667e07:41011 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-12T01:43:33,106 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34189, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-12T01:43:33,107 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:40711, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-12T01:43:33,108 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41417 {}] master.ServerManager(363): Checking decommissioned status of RegionServer b85a53667e07,43203,1733967812588 2024-12-12T01:43:33,108 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41417 {}] master.ServerManager(517): Registering regionserver=b85a53667e07,43203,1733967812588 2024-12-12T01:43:33,108 DEBUG [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca 2024-12-12T01:43:33,108 DEBUG [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:44367 2024-12-12T01:43:33,108 DEBUG [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-12T01:43:33,109 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41417 {}] master.ServerManager(363): Checking decommissioned status of RegionServer b85a53667e07,41011,1733967812624 2024-12-12T01:43:33,110 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41417 {}] master.ServerManager(517): Registering regionserver=b85a53667e07,41011,1733967812624 2024-12-12T01:43:33,110 DEBUG [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca 2024-12-12T01:43:33,110 DEBUG [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:44367 2024-12-12T01:43:33,110 DEBUG [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-12T01:43:33,111 DEBUG [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca 2024-12-12T01:43:33,111 DEBUG [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:44367 2024-12-12T01:43:33,111 DEBUG [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-12T01:43:33,118 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-12T01:43:33,144 DEBUG [RS:0;b85a53667e07:41049 {}] zookeeper.ZKUtil(111): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,144 WARN [RS:0;b85a53667e07:41049 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-12T01:43:33,145 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [b85a53667e07,41049,1733967812549] 2024-12-12T01:43:33,145 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [b85a53667e07,43203,1733967812588] 2024-12-12T01:43:33,145 INFO [RS:0;b85a53667e07:41049 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-12T01:43:33,145 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [b85a53667e07,41011,1733967812624] 2024-12-12T01:43:33,145 DEBUG [RS:1;b85a53667e07:43203 {}] zookeeper.ZKUtil(111): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/b85a53667e07,43203,1733967812588 2024-12-12T01:43:33,145 DEBUG [RS:2;b85a53667e07:41011 {}] zookeeper.ZKUtil(111): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/b85a53667e07,41011,1733967812624 2024-12-12T01:43:33,145 DEBUG [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,145 WARN [RS:2;b85a53667e07:41011 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-12T01:43:33,145 WARN [RS:1;b85a53667e07:43203 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-12T01:43:33,145 INFO [RS:2;b85a53667e07:41011 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-12T01:43:33,145 INFO [RS:1;b85a53667e07:43203 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-12T01:43:33,145 DEBUG [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,41011,1733967812624 2024-12-12T01:43:33,145 DEBUG [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,43203,1733967812588 2024-12-12T01:43:33,152 INFO [RS:0;b85a53667e07:41049 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-12T01:43:33,152 INFO [RS:1;b85a53667e07:43203 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-12T01:43:33,152 INFO [RS:2;b85a53667e07:41011 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-12T01:43:33,155 INFO [RS:1;b85a53667e07:43203 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-12T01:43:33,156 INFO [RS:2;b85a53667e07:41011 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-12T01:43:33,156 INFO [RS:1;b85a53667e07:43203 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-12T01:43:33,156 INFO [RS:2;b85a53667e07:41011 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-12T01:43:33,156 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,156 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,156 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-12T01:43:33,157 INFO [RS:0;b85a53667e07:41049 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-12T01:43:33,157 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-12T01:43:33,158 INFO [RS:2;b85a53667e07:41011 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-12T01:43:33,158 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,158 INFO [RS:0;b85a53667e07:41049 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-12T01:43:33,158 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/b85a53667e07:0, corePoolSize=2, maxPoolSize=2 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,158 INFO [RS:1;b85a53667e07:43203 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,158 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,158 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:2;b85a53667e07:41011 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/b85a53667e07:0, corePoolSize=2, maxPoolSize=2 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:33,159 DEBUG [RS:1;b85a53667e07:43203 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:33,162 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-12T01:43:33,162 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:0;b85a53667e07:41049 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-12T01:43:33,163 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,163 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41011,1733967812624-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-12T01:43:33,163 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,163 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,43203,1733967812588-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/b85a53667e07:0, corePoolSize=2, maxPoolSize=2 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,163 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,164 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,164 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/b85a53667e07:0, corePoolSize=1, maxPoolSize=1 2024-12-12T01:43:33,164 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:33,164 DEBUG [RS:0;b85a53667e07:41049 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0, corePoolSize=3, maxPoolSize=3 2024-12-12T01:43:33,164 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,164 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,164 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,164 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,164 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,164 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41049,1733967812549-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-12T01:43:33,176 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HeapMemoryManager(213): Starting, tuneOn=false 2024-12-12T01:43:33,176 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HeapMemoryManager(213): Starting, tuneOn=false 2024-12-12T01:43:33,176 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HeapMemoryManager(213): Starting, tuneOn=false 2024-12-12T01:43:33,176 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41049,1733967812549-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,176 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41011,1733967812624-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,176 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,43203,1733967812588-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,176 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,176 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,176 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,177 INFO [RS:0;b85a53667e07:41049 {}] regionserver.Replication(171): b85a53667e07,41049,1733967812549 started 2024-12-12T01:43:33,177 INFO [RS:1;b85a53667e07:43203 {}] regionserver.Replication(171): b85a53667e07,43203,1733967812588 started 2024-12-12T01:43:33,177 INFO [RS:2;b85a53667e07:41011 {}] regionserver.Replication(171): b85a53667e07,41011,1733967812624 started 2024-12-12T01:43:33,182 WARN [b85a53667e07:41417 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-12T01:43:33,189 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,189 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,189 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,189 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1482): Serving as b85a53667e07,43203,1733967812588, RpcServer on b85a53667e07/172.17.0.2:43203, sessionid=0x10017bed5b90002 2024-12-12T01:43:33,189 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1482): Serving as b85a53667e07,41049,1733967812549, RpcServer on b85a53667e07/172.17.0.2:41049, sessionid=0x10017bed5b90001 2024-12-12T01:43:33,189 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(1482): Serving as b85a53667e07,41011,1733967812624, RpcServer on b85a53667e07/172.17.0.2:41011, sessionid=0x10017bed5b90003 2024-12-12T01:43:33,189 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-12T01:43:33,189 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-12T01:43:33,189 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-12T01:43:33,189 DEBUG [RS:1;b85a53667e07:43203 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager b85a53667e07,43203,1733967812588 2024-12-12T01:43:33,189 DEBUG [RS:2;b85a53667e07:41011 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager b85a53667e07,41011,1733967812624 2024-12-12T01:43:33,189 DEBUG [RS:0;b85a53667e07:41049 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,189 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,43203,1733967812588' 2024-12-12T01:43:33,189 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,41011,1733967812624' 2024-12-12T01:43:33,189 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,41049,1733967812549' 2024-12-12T01:43:33,189 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-12T01:43:33,189 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-12T01:43:33,189 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-12T01:43:33,190 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-12T01:43:33,190 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-12T01:43:33,190 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-12T01:43:33,190 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-12T01:43:33,190 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-12T01:43:33,190 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-12T01:43:33,190 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-12T01:43:33,190 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-12T01:43:33,190 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-12T01:43:33,190 DEBUG [RS:0;b85a53667e07:41049 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,190 DEBUG [RS:1;b85a53667e07:43203 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager b85a53667e07,43203,1733967812588 2024-12-12T01:43:33,190 DEBUG [RS:2;b85a53667e07:41011 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager b85a53667e07,41011,1733967812624 2024-12-12T01:43:33,191 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,43203,1733967812588' 2024-12-12T01:43:33,191 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,41011,1733967812624' 2024-12-12T01:43:33,191 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'b85a53667e07,41049,1733967812549' 2024-12-12T01:43:33,191 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-12T01:43:33,191 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-12T01:43:33,191 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-12T01:43:33,191 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-12T01:43:33,191 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-12T01:43:33,191 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-12T01:43:33,192 DEBUG [RS:1;b85a53667e07:43203 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-12T01:43:33,192 DEBUG [RS:0;b85a53667e07:41049 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-12T01:43:33,192 INFO [RS:1;b85a53667e07:43203 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-12T01:43:33,192 INFO [RS:1;b85a53667e07:43203 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-12T01:43:33,192 INFO [RS:0;b85a53667e07:41049 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-12T01:43:33,192 DEBUG [RS:2;b85a53667e07:41011 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-12T01:43:33,192 INFO [RS:0;b85a53667e07:41049 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-12T01:43:33,192 INFO [RS:2;b85a53667e07:41011 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-12T01:43:33,192 INFO [RS:2;b85a53667e07:41011 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-12T01:43:33,296 INFO [RS:1;b85a53667e07:43203 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C43203%2C1733967812588, suffix=, logDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,43203,1733967812588, archiveDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/oldWALs, maxLogs=32 2024-12-12T01:43:33,297 INFO [RS:2;b85a53667e07:41011 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C41011%2C1733967812624, suffix=, logDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,41011,1733967812624, archiveDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/oldWALs, maxLogs=32 2024-12-12T01:43:33,297 INFO [RS:0;b85a53667e07:41049 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C41049%2C1733967812549, suffix=, logDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,41049,1733967812549, archiveDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/oldWALs, maxLogs=32 2024-12-12T01:43:33,301 INFO [RS:0;b85a53667e07:41049 {}] monitor.StreamSlowMonitor(122): New stream slow monitor b85a53667e07%2C41049%2C1733967812549.1733967813300 2024-12-12T01:43:33,301 INFO [RS:2;b85a53667e07:41011 {}] monitor.StreamSlowMonitor(122): New stream slow monitor b85a53667e07%2C41011%2C1733967812624.1733967813301 2024-12-12T01:43:33,301 INFO [RS:1;b85a53667e07:43203 {}] monitor.StreamSlowMonitor(122): New stream slow monitor b85a53667e07%2C43203%2C1733967812588.1733967813301 2024-12-12T01:43:33,311 INFO [RS:0;b85a53667e07:41049 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,41049,1733967812549/b85a53667e07%2C41049%2C1733967812549.1733967813300 2024-12-12T01:43:33,311 INFO [RS:2;b85a53667e07:41011 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,41011,1733967812624/b85a53667e07%2C41011%2C1733967812624.1733967813301 2024-12-12T01:43:33,311 INFO [RS:1;b85a53667e07:43203 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,43203,1733967812588/b85a53667e07%2C43203%2C1733967812588.1733967813301 2024-12-12T01:43:33,318 DEBUG [RS:0;b85a53667e07:41049 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38777:38777),(127.0.0.1/127.0.0.1:39021:39021),(127.0.0.1/127.0.0.1:41387:41387)] 2024-12-12T01:43:33,318 DEBUG [RS:2;b85a53667e07:41011 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39021:39021),(127.0.0.1/127.0.0.1:41387:41387),(127.0.0.1/127.0.0.1:38777:38777)] 2024-12-12T01:43:33,318 DEBUG [RS:1;b85a53667e07:43203 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41387:41387),(127.0.0.1/127.0.0.1:39021:39021),(127.0.0.1/127.0.0.1:38777:38777)] 2024-12-12T01:43:33,432 DEBUG [b85a53667e07:41417 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-12T01:43:33,433 DEBUG [b85a53667e07:41417 {}] balancer.BalancerClusterState(204): Hosts are {b85a53667e07=0} racks are {/default-rack=0} 2024-12-12T01:43:33,437 DEBUG [b85a53667e07:41417 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-12T01:43:33,437 DEBUG [b85a53667e07:41417 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-12T01:43:33,437 DEBUG [b85a53667e07:41417 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-12T01:43:33,437 DEBUG [b85a53667e07:41417 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-12T01:43:33,438 DEBUG [b85a53667e07:41417 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-12T01:43:33,438 DEBUG [b85a53667e07:41417 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-12T01:43:33,438 INFO [b85a53667e07:41417 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-12T01:43:33,438 INFO [b85a53667e07:41417 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-12T01:43:33,438 INFO [b85a53667e07:41417 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-12T01:43:33,438 DEBUG [b85a53667e07:41417 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-12T01:43:33,439 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,440 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as b85a53667e07,41049,1733967812549, state=OPENING 2024-12-12T01:43:33,485 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-12T01:43:33,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:33,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:33,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:33,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:33,495 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-12T01:43:33,495 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:33,495 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:33,495 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:33,495 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=b85a53667e07,41049,1733967812549}] 2024-12-12T01:43:33,495 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:33,567 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-12T01:43:33,576 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-12T01:43:33,612 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-12T01:43:33,612 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-12T01:43:33,649 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-12T01:43:33,651 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:38601, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-12T01:43:33,658 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-12T01:43:33,659 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-12T01:43:33,661 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=b85a53667e07%2C41049%2C1733967812549.meta, suffix=.meta, logDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,41049,1733967812549, archiveDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/oldWALs, maxLogs=32 2024-12-12T01:43:33,662 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor b85a53667e07%2C41049%2C1733967812549.meta.1733967813662.meta 2024-12-12T01:43:33,670 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/WALs/b85a53667e07,41049,1733967812549/b85a53667e07%2C41049%2C1733967812549.meta.1733967813662.meta 2024-12-12T01:43:33,677 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38777:38777),(127.0.0.1/127.0.0.1:41387:41387),(127.0.0.1/127.0.0.1:39021:39021)] 2024-12-12T01:43:33,678 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-12T01:43:33,679 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-12T01:43:33,679 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-12T01:43:33,679 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-12T01:43:33,679 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-12T01:43:33,679 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:33,679 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-12T01:43:33,679 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-12T01:43:33,681 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-12T01:43:33,682 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-12T01:43:33,683 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,683 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:33,683 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-12T01:43:33,684 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-12T01:43:33,684 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,685 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:33,685 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-12T01:43:33,686 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-12T01:43:33,686 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,686 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:33,686 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-12T01:43:33,687 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-12T01:43:33,687 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,687 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-12T01:43:33,687 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-12T01:43:33,688 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740 2024-12-12T01:43:33,690 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740 2024-12-12T01:43:33,691 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-12T01:43:33,691 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-12T01:43:33,692 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-12T01:43:33,693 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-12T01:43:33,694 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69647004, jitterRate=0.03782123327255249}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-12T01:43:33,694 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-12T01:43:33,695 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733967813679Writing region info on filesystem at 1733967813679Initializing all the Stores at 1733967813681 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967813681Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967813681Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967813681Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733967813681Cleaning up temporary data from old regions at 1733967813691 (+10 ms)Running coprocessor post-open hooks at 1733967813694 (+3 ms)Region opened successfully at 1733967813695 (+1 ms) 2024-12-12T01:43:33,697 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733967813648 2024-12-12T01:43:33,700 DEBUG [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-12T01:43:33,700 INFO [RS_OPEN_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-12T01:43:33,701 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,703 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as b85a53667e07,41049,1733967812549, state=OPEN 2024-12-12T01:43:33,734 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-12T01:43:33,734 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-12T01:43:33,734 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-12T01:43:33,734 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-12T01:43:33,734 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=b85a53667e07,41049,1733967812549 2024-12-12T01:43:33,734 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:33,734 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:33,734 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:33,734 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-12T01:43:33,740 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-12T01:43:33,740 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=b85a53667e07,41049,1733967812549 in 239 msec 2024-12-12T01:43:33,745 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-12T01:43:33,745 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 712 msec 2024-12-12T01:43:33,746 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-12T01:43:33,746 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-12T01:43:33,747 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-12T01:43:33,747 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=b85a53667e07,41049,1733967812549, seqNum=-1] 2024-12-12T01:43:33,748 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-12T01:43:33,749 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:53765, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-12T01:43:33,757 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 781 msec 2024-12-12T01:43:33,757 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733967813757, completionTime=-1 2024-12-12T01:43:33,757 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-12T01:43:33,758 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-12T01:43:33,761 INFO [master/b85a53667e07:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-12T01:43:33,761 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733967873761 2024-12-12T01:43:33,761 INFO [master/b85a53667e07:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733967933761 2024-12-12T01:43:33,761 INFO [master/b85a53667e07:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 3 msec 2024-12-12T01:43:33,761 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-12T01:43:33,762 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41417,1733967812398-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,762 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41417,1733967812398-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,762 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41417,1733967812398-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,762 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-b85a53667e07:41417, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,762 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,763 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,766 DEBUG [master/b85a53667e07:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-12T01:43:33,769 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.092sec 2024-12-12T01:43:33,770 INFO [master/b85a53667e07:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-12T01:43:33,770 INFO [master/b85a53667e07:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-12T01:43:33,770 INFO [master/b85a53667e07:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-12T01:43:33,770 INFO [master/b85a53667e07:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-12T01:43:33,770 INFO [master/b85a53667e07:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-12T01:43:33,770 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41417,1733967812398-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-12T01:43:33,770 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41417,1733967812398-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-12T01:43:33,773 DEBUG [master/b85a53667e07:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-12T01:43:33,774 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-12T01:43:33,774 INFO [master/b85a53667e07:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=b85a53667e07,41417,1733967812398-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-12T01:43:33,843 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1fdae76f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-12T01:43:33,843 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request b85a53667e07,41417,-1 for getting cluster id 2024-12-12T01:43:33,844 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-12T01:43:33,845 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '45f7c04c-dfa0-431f-9efc-18ac0abe519f' 2024-12-12T01:43:33,845 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-12T01:43:33,845 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "45f7c04c-dfa0-431f-9efc-18ac0abe519f" 2024-12-12T01:43:33,846 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7ae6f79, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-12T01:43:33,846 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [b85a53667e07,41417,-1] 2024-12-12T01:43:33,846 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-12T01:43:33,846 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:33,847 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:39778, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-12T01:43:33,849 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5169f600, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-12T01:43:33,849 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-12T01:43:33,850 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=b85a53667e07,41049,1733967812549, seqNum=-1] 2024-12-12T01:43:33,850 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-12T01:43:33,852 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:47230, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-12T01:43:33,855 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=b85a53667e07,41417,1733967812398 2024-12-12T01:43:33,855 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-12T01:43:33,856 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is b85a53667e07,41417,1733967812398 2024-12-12T01:43:33,857 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@7c7df770 2024-12-12T01:43:33,857 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-12T01:43:33,858 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:39782, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-12T01:43:33,859 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-12T01:43:33,861 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-12T01:43:33,863 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-12T01:43:33,863 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:33,863 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-12T01:43:33,865 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-12T01:43:33,865 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-12T01:43:33,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741837_1013 (size=392) 2024-12-12T01:43:33,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741837_1013 (size=392) 2024-12-12T01:43:33,875 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741837_1013 (size=392) 2024-12-12T01:43:33,877 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 1a751d15c3e9d5508132079cdef0e66d, NAME => 'TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca 2024-12-12T01:43:33,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741838_1014 (size=51) 2024-12-12T01:43:33,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741838_1014 (size=51) 2024-12-12T01:43:33,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741838_1014 (size=51) 2024-12-12T01:43:33,887 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:33,887 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 1a751d15c3e9d5508132079cdef0e66d, disabling compactions & flushes 2024-12-12T01:43:33,887 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:33,887 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:33,887 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. after waiting 0 ms 2024-12-12T01:43:33,887 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:33,887 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:33,887 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 1a751d15c3e9d5508132079cdef0e66d: Waiting for close lock at 1733967813887Disabling compacts and flushes for region at 1733967813887Disabling writes for close at 1733967813887Writing region close event to WAL at 1733967813887Closed at 1733967813887 2024-12-12T01:43:33,889 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-12T01:43:33,889 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733967813889"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733967813889"}]},"ts":"1733967813889"} 2024-12-12T01:43:33,893 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-12T01:43:33,894 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-12T01:43:33,895 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733967813894"}]},"ts":"1733967813894"} 2024-12-12T01:43:33,898 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-12T01:43:33,898 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {b85a53667e07=0} racks are {/default-rack=0} 2024-12-12T01:43:33,899 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-12T01:43:33,899 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-12T01:43:33,900 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-12T01:43:33,900 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-12T01:43:33,900 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-12T01:43:33,900 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-12T01:43:33,900 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-12T01:43:33,900 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-12T01:43:33,900 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-12T01:43:33,900 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-12T01:43:33,900 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=1a751d15c3e9d5508132079cdef0e66d, ASSIGN}] 2024-12-12T01:43:33,902 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=1a751d15c3e9d5508132079cdef0e66d, ASSIGN 2024-12-12T01:43:33,904 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=1a751d15c3e9d5508132079cdef0e66d, ASSIGN; state=OFFLINE, location=b85a53667e07,43203,1733967812588; forceNewPlan=false, retain=false 2024-12-12T01:43:33,972 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-12T01:43:34,054 INFO [b85a53667e07:41417 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-12T01:43:34,055 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=1a751d15c3e9d5508132079cdef0e66d, regionState=OPENING, regionLocation=b85a53667e07,43203,1733967812588 2024-12-12T01:43:34,061 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=1a751d15c3e9d5508132079cdef0e66d, ASSIGN because future has completed 2024-12-12T01:43:34,062 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1a751d15c3e9d5508132079cdef0e66d, server=b85a53667e07,43203,1733967812588}] 2024-12-12T01:43:34,183 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-12T01:43:34,220 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-12T01:43:34,223 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:40985, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-12T01:43:34,227 INFO [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,227 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 1a751d15c3e9d5508132079cdef0e66d, NAME => 'TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d.', STARTKEY => '', ENDKEY => ''} 2024-12-12T01:43:34,228 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,228 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-12T01:43:34,228 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,228 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,230 INFO [StoreOpener-1a751d15c3e9d5508132079cdef0e66d-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,231 INFO [StoreOpener-1a751d15c3e9d5508132079cdef0e66d-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1a751d15c3e9d5508132079cdef0e66d columnFamilyName cf 2024-12-12T01:43:34,231 DEBUG [StoreOpener-1a751d15c3e9d5508132079cdef0e66d-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-12T01:43:34,232 INFO [StoreOpener-1a751d15c3e9d5508132079cdef0e66d-1 {}] regionserver.HStore(327): Store=1a751d15c3e9d5508132079cdef0e66d/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-12T01:43:34,232 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,233 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,233 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,234 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,234 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,235 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,238 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-12T01:43:34,239 INFO [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 1a751d15c3e9d5508132079cdef0e66d; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73929731, jitterRate=0.10163883864879608}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-12T01:43:34,239 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,240 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 1a751d15c3e9d5508132079cdef0e66d: Running coprocessor pre-open hook at 1733967814228Writing region info on filesystem at 1733967814228Initializing all the Stores at 1733967814229 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733967814229Cleaning up temporary data from old regions at 1733967814234 (+5 ms)Running coprocessor post-open hooks at 1733967814239 (+5 ms)Region opened successfully at 1733967814239 2024-12-12T01:43:34,241 INFO [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d., pid=6, masterSystemTime=1733967814219 2024-12-12T01:43:34,244 DEBUG [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,244 INFO [RS_OPEN_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,245 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=1a751d15c3e9d5508132079cdef0e66d, regionState=OPEN, openSeqNum=2, regionLocation=b85a53667e07,43203,1733967812588 2024-12-12T01:43:34,249 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1a751d15c3e9d5508132079cdef0e66d, server=b85a53667e07,43203,1733967812588 because future has completed 2024-12-12T01:43:34,255 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-12T01:43:34,255 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 1a751d15c3e9d5508132079cdef0e66d, server=b85a53667e07,43203,1733967812588 in 189 msec 2024-12-12T01:43:34,259 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-12T01:43:34,260 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=1a751d15c3e9d5508132079cdef0e66d, ASSIGN in 355 msec 2024-12-12T01:43:34,261 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-12T01:43:34,261 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733967814261"}]},"ts":"1733967814261"} 2024-12-12T01:43:34,264 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-12T01:43:34,266 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-12T01:43:34,269 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 407 msec 2024-12-12T01:43:34,492 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-12T01:43:34,493 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-12T01:43:34,493 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-12T01:43:34,493 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-12T01:43:34,499 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-12T01:43:34,499 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-12T01:43:34,499 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-12T01:43:34,504 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d., hostname=b85a53667e07,43203,1733967812588, seqNum=2] 2024-12-12T01:43:34,504 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-12T01:43:34,506 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46800, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-12T01:43:34,509 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-12T01:43:34,510 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-12T01:43:34,512 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-12T01:43:34,512 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-12T01:43:34,513 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-12T01:43:34,513 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-12T01:43:34,622 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-12T01:43:34,667 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=43203 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-12T01:43:34,668 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,668 INFO [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 1a751d15c3e9d5508132079cdef0e66d 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-12T01:43:34,683 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d/.tmp/cf/cb7f0d93b50f4f2ab660b73a8c2a137d is 36, key is row/cf:cq/1733967814506/Put/seqid=0 2024-12-12T01:43:34,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741839_1015 (size=4787) 2024-12-12T01:43:34,690 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741839_1015 (size=4787) 2024-12-12T01:43:34,690 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741839_1015 (size=4787) 2024-12-12T01:43:34,691 INFO [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d/.tmp/cf/cb7f0d93b50f4f2ab660b73a8c2a137d 2024-12-12T01:43:34,701 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d/.tmp/cf/cb7f0d93b50f4f2ab660b73a8c2a137d as hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d/cf/cb7f0d93b50f4f2ab660b73a8c2a137d 2024-12-12T01:43:34,709 INFO [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d/cf/cb7f0d93b50f4f2ab660b73a8c2a137d, entries=1, sequenceid=5, filesize=4.7 K 2024-12-12T01:43:34,711 INFO [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 1a751d15c3e9d5508132079cdef0e66d in 43ms, sequenceid=5, compaction requested=false 2024-12-12T01:43:34,711 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 1a751d15c3e9d5508132079cdef0e66d: 2024-12-12T01:43:34,711 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,711 DEBUG [RS_FLUSH_OPERATIONS-regionserver/b85a53667e07:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-12T01:43:34,712 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-12T01:43:34,717 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-12T01:43:34,717 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 201 msec 2024-12-12T01:43:34,721 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 209 msec 2024-12-12T01:43:34,833 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41417 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-12T01:43:34,834 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-12T01:43:34,842 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-12T01:43:34,842 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-12T01:43:34,842 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:34,842 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:34,843 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:34,843 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-12T01:43:34,843 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-12T01:43:34,843 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=103898131, stopped=false 2024-12-12T01:43:34,844 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=b85a53667e07,41417,1733967812398 2024-12-12T01:43:34,876 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:34,876 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:34,876 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:34,876 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-12T01:43:34,876 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:34,876 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:34,876 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:34,876 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:34,876 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-12T01:43:34,877 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-12T01:43:34,877 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:34,877 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:34,878 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:34,878 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:34,878 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'b85a53667e07,41049,1733967812549' ***** 2024-12-12T01:43:34,878 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-12T01:43:34,878 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'b85a53667e07,43203,1733967812588' ***** 2024-12-12T01:43:34,878 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-12T01:43:34,878 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'b85a53667e07,41011,1733967812624' ***** 2024-12-12T01:43:34,878 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HeapMemoryManager(224): Stopping 2024-12-12T01:43:34,878 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-12T01:43:34,879 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:34,879 INFO [RS:0;b85a53667e07:41049 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-12T01:43:34,879 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HeapMemoryManager(224): Stopping 2024-12-12T01:43:34,879 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-12T01:43:34,879 INFO [RS:0;b85a53667e07:41049 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-12T01:43:34,879 INFO [RS:1;b85a53667e07:43203 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-12T01:43:34,880 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(959): stopping server b85a53667e07,41049,1733967812549 2024-12-12T01:43:34,879 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-12T01:43:34,880 INFO [RS:0;b85a53667e07:41049 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-12T01:43:34,879 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-12T01:43:34,880 INFO [RS:1;b85a53667e07:43203 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-12T01:43:34,880 INFO [RS:0;b85a53667e07:41049 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;b85a53667e07:41049. 2024-12-12T01:43:34,880 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(3091): Received CLOSE for 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,880 DEBUG [RS:0;b85a53667e07:41049 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:34,880 DEBUG [RS:0;b85a53667e07:41049 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:34,880 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HeapMemoryManager(224): Stopping 2024-12-12T01:43:34,880 INFO [RS:0;b85a53667e07:41049 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-12T01:43:34,880 INFO [RS:0;b85a53667e07:41049 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-12T01:43:34,880 INFO [RS:2;b85a53667e07:41011 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-12T01:43:34,880 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-12T01:43:34,881 INFO [RS:0;b85a53667e07:41049 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-12T01:43:34,881 INFO [RS:2;b85a53667e07:41011 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-12T01:43:34,881 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(959): stopping server b85a53667e07,41011,1733967812624 2024-12-12T01:43:34,881 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-12T01:43:34,881 INFO [RS:2;b85a53667e07:41011 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-12T01:43:34,881 INFO [RS:2;b85a53667e07:41011 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;b85a53667e07:41011. 2024-12-12T01:43:34,881 DEBUG [RS:2;b85a53667e07:41011 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:34,881 DEBUG [RS:2;b85a53667e07:41011 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:34,881 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(976): stopping server b85a53667e07,41011,1733967812624; all regions closed. 2024-12-12T01:43:34,882 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(959): stopping server b85a53667e07,43203,1733967812588 2024-12-12T01:43:34,882 INFO [RS:1;b85a53667e07:43203 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-12T01:43:34,882 INFO [RS:1;b85a53667e07:43203 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;b85a53667e07:43203. 2024-12-12T01:43:34,882 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 1a751d15c3e9d5508132079cdef0e66d, disabling compactions & flushes 2024-12-12T01:43:34,882 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-12T01:43:34,882 INFO [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,882 DEBUG [RS:1;b85a53667e07:43203 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-12T01:43:34,882 DEBUG [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-12T01:43:34,882 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,882 DEBUG [RS:1;b85a53667e07:43203 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:34,882 DEBUG [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-12T01:43:34,882 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. after waiting 0 ms 2024-12-12T01:43:34,882 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-12T01:43:34,883 DEBUG [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1325): Online Regions={1a751d15c3e9d5508132079cdef0e66d=TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d.} 2024-12-12T01:43:34,883 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,883 DEBUG [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1351): Waiting on 1a751d15c3e9d5508132079cdef0e66d 2024-12-12T01:43:34,883 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-12T01:43:34,883 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-12T01:43:34,883 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:34,883 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-12T01:43:34,883 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-12T01:43:34,883 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:34,883 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-12T01:43:34,883 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:34,884 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:34,884 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-12T01:43:34,884 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:34,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741834_1010 (size=93) 2024-12-12T01:43:34,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741834_1010 (size=93) 2024-12-12T01:43:34,888 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741834_1010 (size=93) 2024-12-12T01:43:34,891 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/default/TestHBaseWalOnEC/1a751d15c3e9d5508132079cdef0e66d/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-12T01:43:34,892 DEBUG [RS:2;b85a53667e07:41011 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/oldWALs 2024-12-12T01:43:34,892 INFO [RS:2;b85a53667e07:41011 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog b85a53667e07%2C41011%2C1733967812624:(num 1733967813301) 2024-12-12T01:43:34,892 DEBUG [RS:2;b85a53667e07:41011 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:34,892 INFO [RS:2;b85a53667e07:41011 {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:34,892 INFO [RS:2;b85a53667e07:41011 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-12T01:43:34,892 INFO [RS:2;b85a53667e07:41011 {}] hbase.ChoreService(370): Chore service for: regionserver/b85a53667e07:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-12T01:43:34,892 INFO [RS:2;b85a53667e07:41011 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-12T01:43:34,892 INFO [regionserver/b85a53667e07:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-12T01:43:34,892 INFO [RS:2;b85a53667e07:41011 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-12T01:43:34,892 INFO [RS:2;b85a53667e07:41011 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-12T01:43:34,892 INFO [RS:2;b85a53667e07:41011 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-12T01:43:34,893 INFO [RS:2;b85a53667e07:41011 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41011 2024-12-12T01:43:34,893 INFO [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,893 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 1a751d15c3e9d5508132079cdef0e66d: Waiting for close lock at 1733967814882Running coprocessor pre-close hooks at 1733967814882Disabling compacts and flushes for region at 1733967814882Disabling writes for close at 1733967814882Writing region close event to WAL at 1733967814884 (+2 ms)Running coprocessor post-close hooks at 1733967814892 (+8 ms)Closed at 1733967814893 (+1 ms) 2024-12-12T01:43:34,893 DEBUG [RS_CLOSE_REGION-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d. 2024-12-12T01:43:34,903 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/info/699dae14a56a4dd5b648286251771d66 is 153, key is TestHBaseWalOnEC,,1733967813859.1a751d15c3e9d5508132079cdef0e66d./info:regioninfo/1733967814245/Put/seqid=0 2024-12-12T01:43:34,909 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741840_1016 (size=6637) 2024-12-12T01:43:34,909 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741840_1016 (size=6637) 2024-12-12T01:43:34,909 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741840_1016 (size=6637) 2024-12-12T01:43:34,910 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/info/699dae14a56a4dd5b648286251771d66 2024-12-12T01:43:34,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/b85a53667e07,41011,1733967812624 2024-12-12T01:43:34,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-12T01:43:34,926 INFO [RS:2;b85a53667e07:41011 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-12T01:43:34,934 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [b85a53667e07,41011,1733967812624] 2024-12-12T01:43:34,937 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/ns/d6264b7b594a4a8698e6aa72a6fc1a89 is 43, key is default/ns:d/1733967813750/Put/seqid=0 2024-12-12T01:43:34,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741841_1017 (size=5153) 2024-12-12T01:43:34,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741841_1017 (size=5153) 2024-12-12T01:43:34,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741841_1017 (size=5153) 2024-12-12T01:43:34,944 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/b85a53667e07,41011,1733967812624 already deleted, retry=false 2024-12-12T01:43:34,944 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; b85a53667e07,41011,1733967812624 expired; onlineServers=2 2024-12-12T01:43:34,944 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/ns/d6264b7b594a4a8698e6aa72a6fc1a89 2024-12-12T01:43:34,965 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/table/d22b7fb7db794833bace5ec85f303096 is 52, key is TestHBaseWalOnEC/table:state/1733967814261/Put/seqid=0 2024-12-12T01:43:34,965 INFO [regionserver/b85a53667e07:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:34,965 INFO [regionserver/b85a53667e07:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:34,966 INFO [regionserver/b85a53667e07:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:34,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741842_1018 (size=5249) 2024-12-12T01:43:34,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741842_1018 (size=5249) 2024-12-12T01:43:34,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741842_1018 (size=5249) 2024-12-12T01:43:34,972 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/table/d22b7fb7db794833bace5ec85f303096 2024-12-12T01:43:34,979 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/info/699dae14a56a4dd5b648286251771d66 as hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/info/699dae14a56a4dd5b648286251771d66 2024-12-12T01:43:34,987 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/info/699dae14a56a4dd5b648286251771d66, entries=10, sequenceid=11, filesize=6.5 K 2024-12-12T01:43:34,988 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/ns/d6264b7b594a4a8698e6aa72a6fc1a89 as hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/ns/d6264b7b594a4a8698e6aa72a6fc1a89 2024-12-12T01:43:34,995 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/ns/d6264b7b594a4a8698e6aa72a6fc1a89, entries=2, sequenceid=11, filesize=5.0 K 2024-12-12T01:43:34,996 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/.tmp/table/d22b7fb7db794833bace5ec85f303096 as hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/table/d22b7fb7db794833bace5ec85f303096 2024-12-12T01:43:35,003 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/table/d22b7fb7db794833bace5ec85f303096, entries=2, sequenceid=11, filesize=5.1 K 2024-12-12T01:43:35,005 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 121ms, sequenceid=11, compaction requested=false 2024-12-12T01:43:35,010 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-12T01:43:35,010 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-12T01:43:35,010 INFO [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-12T01:43:35,011 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733967814883Running coprocessor pre-close hooks at 1733967814883Disabling compacts and flushes for region at 1733967814883Disabling writes for close at 1733967814883Obtaining lock to block concurrent updates at 1733967814884 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1733967814884Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733967814884Flushing stores of hbase:meta,,1.1588230740 at 1733967814885 (+1 ms)Flushing 1588230740/info: creating writer at 1733967814885Flushing 1588230740/info: appending metadata at 1733967814902 (+17 ms)Flushing 1588230740/info: closing flushed file at 1733967814903 (+1 ms)Flushing 1588230740/ns: creating writer at 1733967814918 (+15 ms)Flushing 1588230740/ns: appending metadata at 1733967814936 (+18 ms)Flushing 1588230740/ns: closing flushed file at 1733967814936Flushing 1588230740/table: creating writer at 1733967814950 (+14 ms)Flushing 1588230740/table: appending metadata at 1733967814964 (+14 ms)Flushing 1588230740/table: closing flushed file at 1733967814964Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4969086c: reopening flushed file at 1733967814978 (+14 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@49e73284: reopening flushed file at 1733967814987 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7b0f00ee: reopening flushed file at 1733967814995 (+8 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 121ms, sequenceid=11, compaction requested=false at 1733967815005 (+10 ms)Writing region close event to WAL at 1733967815006 (+1 ms)Running coprocessor post-close hooks at 1733967815010 (+4 ms)Closed at 1733967815010 2024-12-12T01:43:35,011 DEBUG [RS_CLOSE_META-regionserver/b85a53667e07:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-12T01:43:35,035 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:35,035 INFO [RS:2;b85a53667e07:41011 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-12T01:43:35,035 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41011-0x10017bed5b90003, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:35,035 INFO [RS:2;b85a53667e07:41011 {}] regionserver.HRegionServer(1031): Exiting; stopping=b85a53667e07,41011,1733967812624; zookeeper connection closed. 2024-12-12T01:43:35,035 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4c8aa3b6 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4c8aa3b6 2024-12-12T01:43:35,083 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(976): stopping server b85a53667e07,41049,1733967812549; all regions closed. 2024-12-12T01:43:35,083 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(976): stopping server b85a53667e07,43203,1733967812588; all regions closed. 2024-12-12T01:43:35,083 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,083 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,083 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,084 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,084 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,084 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,084 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,084 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,084 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,084 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,087 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741836_1012 (size=2751) 2024-12-12T01:43:35,087 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741835_1011 (size=1298) 2024-12-12T01:43:35,087 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741836_1012 (size=2751) 2024-12-12T01:43:35,087 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741836_1012 (size=2751) 2024-12-12T01:43:35,089 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741835_1011 (size=1298) 2024-12-12T01:43:35,089 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741835_1011 (size=1298) 2024-12-12T01:43:35,090 DEBUG [RS:1;b85a53667e07:43203 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/oldWALs 2024-12-12T01:43:35,091 DEBUG [RS:0;b85a53667e07:41049 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/oldWALs 2024-12-12T01:43:35,091 INFO [RS:1;b85a53667e07:43203 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog b85a53667e07%2C43203%2C1733967812588:(num 1733967813301) 2024-12-12T01:43:35,091 INFO [RS:0;b85a53667e07:41049 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog b85a53667e07%2C41049%2C1733967812549.meta:.meta(num 1733967813662) 2024-12-12T01:43:35,091 DEBUG [RS:1;b85a53667e07:43203 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:35,091 INFO [RS:1;b85a53667e07:43203 {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:35,091 INFO [RS:1;b85a53667e07:43203 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-12T01:43:35,091 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,091 INFO [RS:1;b85a53667e07:43203 {}] hbase.ChoreService(370): Chore service for: regionserver/b85a53667e07:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-12T01:43:35,091 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,091 INFO [RS:1;b85a53667e07:43203 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-12T01:43:35,091 INFO [regionserver/b85a53667e07:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-12T01:43:35,091 INFO [RS:1;b85a53667e07:43203 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-12T01:43:35,091 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,091 INFO [RS:1;b85a53667e07:43203 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-12T01:43:35,091 INFO [RS:1;b85a53667e07:43203 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-12T01:43:35,091 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,092 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,092 INFO [RS:1;b85a53667e07:43203 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:43203 2024-12-12T01:43:35,095 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741833_1009 (size=93) 2024-12-12T01:43:35,095 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741833_1009 (size=93) 2024-12-12T01:43:35,095 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741833_1009 (size=93) 2024-12-12T01:43:35,101 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/b85a53667e07,43203,1733967812588 2024-12-12T01:43:35,101 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-12T01:43:35,101 INFO [RS:1;b85a53667e07:43203 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-12T01:43:35,110 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [b85a53667e07,43203,1733967812588] 2024-12-12T01:43:35,118 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/b85a53667e07,43203,1733967812588 already deleted, retry=false 2024-12-12T01:43:35,118 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; b85a53667e07,43203,1733967812588 expired; onlineServers=1 2024-12-12T01:43:35,191 INFO [regionserver/b85a53667e07:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-12T01:43:35,192 INFO [regionserver/b85a53667e07:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-12T01:43:35,210 INFO [RS:1;b85a53667e07:43203 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-12T01:43:35,210 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:35,210 INFO [RS:1;b85a53667e07:43203 {}] regionserver.HRegionServer(1031): Exiting; stopping=b85a53667e07,43203,1733967812588; zookeeper connection closed. 2024-12-12T01:43:35,210 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43203-0x10017bed5b90002, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:35,210 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@71f71f44 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@71f71f44 2024-12-12T01:43:35,503 DEBUG [RS:0;b85a53667e07:41049 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/oldWALs 2024-12-12T01:43:35,503 INFO [RS:0;b85a53667e07:41049 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog b85a53667e07%2C41049%2C1733967812549:(num 1733967813300) 2024-12-12T01:43:35,503 DEBUG [RS:0;b85a53667e07:41049 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-12T01:43:35,503 INFO [RS:0;b85a53667e07:41049 {}] regionserver.LeaseManager(133): Closed leases 2024-12-12T01:43:35,504 INFO [RS:0;b85a53667e07:41049 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-12T01:43:35,504 INFO [RS:0;b85a53667e07:41049 {}] hbase.ChoreService(370): Chore service for: regionserver/b85a53667e07:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-12T01:43:35,505 INFO [RS:0;b85a53667e07:41049 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-12T01:43:35,505 INFO [regionserver/b85a53667e07:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-12T01:43:35,505 INFO [RS:0;b85a53667e07:41049 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41049 2024-12-12T01:43:35,534 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/b85a53667e07,41049,1733967812549 2024-12-12T01:43:35,535 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-12T01:43:35,535 INFO [RS:0;b85a53667e07:41049 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-12T01:43:35,543 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [b85a53667e07,41049,1733967812549] 2024-12-12T01:43:35,551 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/b85a53667e07,41049,1733967812549 already deleted, retry=false 2024-12-12T01:43:35,551 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; b85a53667e07,41049,1733967812549 expired; onlineServers=0 2024-12-12T01:43:35,551 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'b85a53667e07,41417,1733967812398' ***** 2024-12-12T01:43:35,551 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-12T01:43:35,552 INFO [M:0;b85a53667e07:41417 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-12T01:43:35,552 INFO [M:0;b85a53667e07:41417 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-12T01:43:35,552 DEBUG [M:0;b85a53667e07:41417 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-12T01:43:35,552 DEBUG [M:0;b85a53667e07:41417 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-12T01:43:35,552 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-12T01:43:35,552 DEBUG [master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.small.0-1733967812981 {}] cleaner.HFileCleaner(306): Exit Thread[master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.small.0-1733967812981,5,FailOnTimeoutGroup] 2024-12-12T01:43:35,552 DEBUG [master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.large.0-1733967812981 {}] cleaner.HFileCleaner(306): Exit Thread[master/b85a53667e07:0:becomeActiveMaster-HFileCleaner.large.0-1733967812981,5,FailOnTimeoutGroup] 2024-12-12T01:43:35,553 INFO [M:0;b85a53667e07:41417 {}] hbase.ChoreService(370): Chore service for: master/b85a53667e07:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-12T01:43:35,553 INFO [M:0;b85a53667e07:41417 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-12T01:43:35,554 DEBUG [M:0;b85a53667e07:41417 {}] master.HMaster(1795): Stopping service threads 2024-12-12T01:43:35,554 INFO [M:0;b85a53667e07:41417 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-12T01:43:35,554 INFO [M:0;b85a53667e07:41417 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-12T01:43:35,554 INFO [M:0;b85a53667e07:41417 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-12T01:43:35,555 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-12T01:43:35,560 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-12T01:43:35,560 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-12T01:43:35,560 DEBUG [M:0;b85a53667e07:41417 {}] zookeeper.ZKUtil(347): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-12T01:43:35,560 WARN [M:0;b85a53667e07:41417 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-12T01:43:35,562 INFO [M:0;b85a53667e07:41417 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/.lastflushedseqids 2024-12-12T01:43:35,570 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741843_1019 (size=127) 2024-12-12T01:43:35,570 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741843_1019 (size=127) 2024-12-12T01:43:35,570 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741843_1019 (size=127) 2024-12-12T01:43:35,571 INFO [M:0;b85a53667e07:41417 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-12T01:43:35,571 INFO [M:0;b85a53667e07:41417 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-12T01:43:35,571 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-12T01:43:35,571 INFO [M:0;b85a53667e07:41417 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:35,571 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:35,571 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-12T01:43:35,571 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:35,571 INFO [M:0;b85a53667e07:41417 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.83 KB heapSize=34.12 KB 2024-12-12T01:43:35,588 DEBUG [M:0;b85a53667e07:41417 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/7c56714014bc44a289f50d026636c61d is 82, key is hbase:meta,,1/info:regioninfo/1733967813701/Put/seqid=0 2024-12-12T01:43:35,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741844_1020 (size=5672) 2024-12-12T01:43:35,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741844_1020 (size=5672) 2024-12-12T01:43:35,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741844_1020 (size=5672) 2024-12-12T01:43:35,595 INFO [M:0;b85a53667e07:41417 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/7c56714014bc44a289f50d026636c61d 2024-12-12T01:43:35,618 DEBUG [M:0;b85a53667e07:41417 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/31084772204248129fdafa4f14b45f37 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733967814268/Put/seqid=0 2024-12-12T01:43:35,621 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-12T01:43:35,621 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-12T01:43:35,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741845_1021 (size=6439) 2024-12-12T01:43:35,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741845_1021 (size=6439) 2024-12-12T01:43:35,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741845_1021 (size=6439) 2024-12-12T01:43:35,626 INFO [M:0;b85a53667e07:41417 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/31084772204248129fdafa4f14b45f37 2024-12-12T01:43:35,643 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:35,643 INFO [RS:0;b85a53667e07:41049 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-12T01:43:35,643 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41049-0x10017bed5b90001, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:35,643 INFO [RS:0;b85a53667e07:41049 {}] regionserver.HRegionServer(1031): Exiting; stopping=b85a53667e07,41049,1733967812549; zookeeper connection closed. 2024-12-12T01:43:35,643 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@2c5fe3d4 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@2c5fe3d4 2024-12-12T01:43:35,644 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-12T01:43:35,645 DEBUG [M:0;b85a53667e07:41417 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c31ad6d73f114f069e8cb35882ac034c is 69, key is b85a53667e07,41011,1733967812624/rs:state/1733967813110/Put/seqid=0 2024-12-12T01:43:35,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741846_1022 (size=5294) 2024-12-12T01:43:35,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741846_1022 (size=5294) 2024-12-12T01:43:35,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741846_1022 (size=5294) 2024-12-12T01:43:35,652 INFO [M:0;b85a53667e07:41417 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c31ad6d73f114f069e8cb35882ac034c 2024-12-12T01:43:35,659 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/7c56714014bc44a289f50d026636c61d as hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/7c56714014bc44a289f50d026636c61d 2024-12-12T01:43:35,665 INFO [M:0;b85a53667e07:41417 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/7c56714014bc44a289f50d026636c61d, entries=8, sequenceid=72, filesize=5.5 K 2024-12-12T01:43:35,666 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/31084772204248129fdafa4f14b45f37 as hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/31084772204248129fdafa4f14b45f37 2024-12-12T01:43:35,672 INFO [M:0;b85a53667e07:41417 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/31084772204248129fdafa4f14b45f37, entries=8, sequenceid=72, filesize=6.3 K 2024-12-12T01:43:35,674 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c31ad6d73f114f069e8cb35882ac034c as hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/c31ad6d73f114f069e8cb35882ac034c 2024-12-12T01:43:35,680 INFO [M:0;b85a53667e07:41417 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:44367/user/jenkins/test-data/5de5a1a5-4531-c6a0-09bc-62d0047522ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/c31ad6d73f114f069e8cb35882ac034c, entries=3, sequenceid=72, filesize=5.2 K 2024-12-12T01:43:35,681 INFO [M:0;b85a53667e07:41417 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 110ms, sequenceid=72, compaction requested=false 2024-12-12T01:43:35,683 INFO [M:0;b85a53667e07:41417 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-12T01:43:35,683 DEBUG [M:0;b85a53667e07:41417 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733967815571Disabling compacts and flushes for region at 1733967815571Disabling writes for close at 1733967815571Obtaining lock to block concurrent updates at 1733967815571Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733967815571Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27471, getHeapSize=34872, getOffHeapSize=0, getCellsCount=85 at 1733967815572 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733967815573 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733967815573Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733967815588 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733967815588Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733967815601 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733967815618 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733967815618Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733967815631 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733967815645 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733967815645Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@47b0a583: reopening flushed file at 1733967815657 (+12 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@404036f1: reopening flushed file at 1733967815665 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@711f9d8b: reopening flushed file at 1733967815673 (+8 ms)Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 110ms, sequenceid=72, compaction requested=false at 1733967815681 (+8 ms)Writing region close event to WAL at 1733967815683 (+2 ms)Closed at 1733967815683 2024-12-12T01:43:35,686 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,686 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,687 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,687 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,687 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-12T01:43:35,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39739 is added to blk_1073741830_1006 (size=32674) 2024-12-12T01:43:35,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45293 is added to blk_1073741830_1006 (size=32674) 2024-12-12T01:43:35,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38519 is added to blk_1073741830_1006 (size=32674) 2024-12-12T01:43:35,690 INFO [M:0;b85a53667e07:41417 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-12T01:43:35,690 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-12T01:43:35,690 INFO [M:0;b85a53667e07:41417 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41417 2024-12-12T01:43:35,690 INFO [M:0;b85a53667e07:41417 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-12T01:43:35,817 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:35,817 INFO [M:0;b85a53667e07:41417 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-12T01:43:35,817 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41417-0x10017bed5b90000, quorum=127.0.0.1:63173, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-12T01:43:35,823 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@53d1de4b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:35,823 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@584937f8{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-12T01:43:35,824 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-12T01:43:35,824 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@40bb076c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-12T01:43:35,824 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@74dea45e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir/,STOPPED} 2024-12-12T01:43:35,827 WARN [BP-2131542035-172.17.0.2-1733967810531 heartbeating to localhost/127.0.0.1:44367 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-12T01:43:35,827 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-12T01:43:35,827 WARN [BP-2131542035-172.17.0.2-1733967810531 heartbeating to localhost/127.0.0.1:44367 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-2131542035-172.17.0.2-1733967810531 (Datanode Uuid fcb0eafa-82f6-4c36-9420-5c0bb0b50f5a) service to localhost/127.0.0.1:44367 2024-12-12T01:43:35,827 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-12T01:43:35,829 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data5/current/BP-2131542035-172.17.0.2-1733967810531 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:35,829 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data6/current/BP-2131542035-172.17.0.2-1733967810531 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:35,829 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-12T01:43:35,831 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@31df22e4{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:35,831 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1c756a11{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-12T01:43:35,831 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-12T01:43:35,831 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@24b4498d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-12T01:43:35,832 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5a1d6627{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir/,STOPPED} 2024-12-12T01:43:35,833 WARN [BP-2131542035-172.17.0.2-1733967810531 heartbeating to localhost/127.0.0.1:44367 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-12T01:43:35,833 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-12T01:43:35,833 WARN [BP-2131542035-172.17.0.2-1733967810531 heartbeating to localhost/127.0.0.1:44367 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-2131542035-172.17.0.2-1733967810531 (Datanode Uuid 4caee6f0-92fb-489e-a0bf-c2a9ca308efd) service to localhost/127.0.0.1:44367 2024-12-12T01:43:35,833 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-12T01:43:35,833 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data3/current/BP-2131542035-172.17.0.2-1733967810531 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:35,833 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data4/current/BP-2131542035-172.17.0.2-1733967810531 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:35,834 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-12T01:43:35,835 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@24fac726{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-12T01:43:35,836 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11c2a472{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-12T01:43:35,836 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-12T01:43:35,836 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@56d7456a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-12T01:43:35,836 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3dec3425{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir/,STOPPED} 2024-12-12T01:43:35,837 WARN [BP-2131542035-172.17.0.2-1733967810531 heartbeating to localhost/127.0.0.1:44367 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-12T01:43:35,837 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-12T01:43:35,837 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-12T01:43:35,837 WARN [BP-2131542035-172.17.0.2-1733967810531 heartbeating to localhost/127.0.0.1:44367 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-2131542035-172.17.0.2-1733967810531 (Datanode Uuid f4693875-b11f-4404-9a8e-9e903552c827) service to localhost/127.0.0.1:44367 2024-12-12T01:43:35,838 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data1/current/BP-2131542035-172.17.0.2-1733967810531 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:35,838 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/cluster_bc4050e8-fb6b-f55e-4e42-2be47965612c/data/data2/current/BP-2131542035-172.17.0.2-1733967810531 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-12T01:43:35,838 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-12T01:43:35,844 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@391ced48{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-12T01:43:35,844 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@38b34206{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-12T01:43:35,844 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-12T01:43:35,845 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7940a36d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-12T01:43:35,845 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@42b66dc7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/4c3d4028-fd17-e6ef-19ca-6fc6ab48b86f/hadoop.log.dir/,STOPPED} 2024-12-12T01:43:35,851 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-12T01:43:35,874 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-12T01:43:35,880 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=149 (was 87) - Thread LEAK? -, OpenFileDescriptor=518 (was 439) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=343 (was 329) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=9256 (was 9447)