2023-08-10 06:31:54,390 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@13c9d689 2023-08-10 06:31:54,405 main DEBUG Took 0.013186 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2023-08-10 06:31:54,406 main DEBUG PluginManager 'Core' found 129 plugins 2023-08-10 06:31:54,406 main DEBUG PluginManager 'Level' found 0 plugins 2023-08-10 06:31:54,407 main DEBUG PluginManager 'Lookup' found 16 plugins 2023-08-10 06:31:54,408 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,418 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2023-08-10 06:31:54,435 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,437 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,438 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,438 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,439 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,439 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,440 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,441 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,441 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,442 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,442 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,443 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,443 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,444 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,444 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,445 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,445 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,446 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,446 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,447 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,447 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,447 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,448 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,448 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2023-08-10 06:31:54,449 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,449 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2023-08-10 06:31:54,452 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2023-08-10 06:31:54,454 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2023-08-10 06:31:54,456 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2023-08-10 06:31:54,456 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2023-08-10 06:31:54,457 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2023-08-10 06:31:54,457 main DEBUG PluginManager 'Converter' found 47 plugins 2023-08-10 06:31:54,469 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2023-08-10 06:31:54,471 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2023-08-10 06:31:54,475 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2023-08-10 06:31:54,476 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2023-08-10 06:31:54,476 main DEBUG createAppenders(={Console}) 2023-08-10 06:31:54,477 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@13c9d689 initialized 2023-08-10 06:31:54,477 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@13c9d689 2023-08-10 06:31:54,477 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@13c9d689 OK. 2023-08-10 06:31:54,478 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2023-08-10 06:31:54,478 main DEBUG OutputStream closed 2023-08-10 06:31:54,478 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2023-08-10 06:31:54,479 main DEBUG Appender DefaultConsole-1 stopped with status true 2023-08-10 06:31:54,479 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@15c43bd9 OK 2023-08-10 06:31:54,525 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586 2023-08-10 06:31:54,528 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=StatusLogger 2023-08-10 06:31:54,529 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=ContextSelector 2023-08-10 06:31:54,531 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name= 2023-08-10 06:31:54,531 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.directory 2023-08-10 06:31:54,532 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2023-08-10 06:31:54,532 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.zookeeper 2023-08-10 06:31:54,532 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2023-08-10 06:31:54,533 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2023-08-10 06:31:54,533 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2023-08-10 06:31:54,533 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.hbase 2023-08-10 06:31:54,534 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop 2023-08-10 06:31:54,534 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2023-08-10 06:31:54,534 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2023-08-10 06:31:54,534 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2023-08-10 06:31:54,535 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2023-08-10 06:31:54,535 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2023-08-10 06:31:54,536 main DEBUG Registering MBean org.apache.logging.log4j2:type=1b6d3586,component=Appenders,name=Console 2023-08-10 06:31:54,540 main DEBUG org.apache.logging.log4j.core.util.SystemClock does not support precise timestamps. 2023-08-10 06:31:54,540 main DEBUG Reconfiguration complete for context[name=1b6d3586] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-logging/target/hbase-logging-2.6.0-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@1b6e1eff) with optional ClassLoader: null 2023-08-10 06:31:54,540 main DEBUG Shutdown hook enabled. Registering a new one. 2023-08-10 06:31:54,541 main DEBUG LoggerContext[name=1b6d3586, org.apache.logging.log4j.core.LoggerContext@1b6e1eff] started OK. 2023-08-10T06:31:54,572 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.backup.TestRemoteRestore timeout: 13 mins 2023-08-10 06:31:54,575 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2023-08-10 06:31:54,576 main DEBUG org.apache.logging.log4j.core.util.SystemClock does not support precise timestamps. 2023-08-10T06:31:54,898 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(348): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9 2023-08-10T06:31:54,903 DEBUG [Time-limited test {}] impl.BackupManager(127): Added log cleaner: org.apache.hadoop.hbase.backup.master.BackupLogCleaner. Added master procedure manager: org.apache.hadoop.hbase.backup.master.LogRollMasterProcedureManager.Added master procedure manager: org.apache.hadoop.hbase.backup.BackupHFileCleaner 2023-08-10T06:31:54,905 DEBUG [Time-limited test {}] impl.BackupManager(157): Added region procedure manager: org.apache.hadoop.hbase.backup.regionserver.LogRollRegionServerProcedureManager. Added region observer: org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:31:54,906 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=1, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2023-08-10T06:31:54,909 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/cluster_31ee86f8-f5ed-8a45-aa31-8b420ad4a168, deleteOnExit=true 2023-08-10T06:31:54,909 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2023-08-10T06:31:54,910 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/test.cache.data in system properties and HBase conf 2023-08-10T06:31:54,910 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.tmp.dir in system properties and HBase conf 2023-08-10T06:31:54,911 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.log.dir in system properties and HBase conf 2023-08-10T06:31:54,911 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.local.dir in system properties and HBase conf 2023-08-10T06:31:54,911 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.temp.dir in system properties and HBase conf 2023-08-10T06:31:54,911 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2023-08-10T06:31:55,004 WARN [Time-limited test {}] util.NativeCodeLoader(62): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2023-08-10T06:31:55,428 DEBUG [Time-limited test {}] fs.HFileSystem(308): The file system is not a DistributedFileSystem. Skipping on block location reordering 2023-08-10T06:31:55,433 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:31:55,433 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:31:55,434 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.nodemanager.log-dirs in system properties and HBase conf 2023-08-10T06:31:55,434 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:31:55,435 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2023-08-10T06:31:55,435 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2023-08-10T06:31:55,436 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:31:55,436 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:31:55,437 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2023-08-10T06:31:55,437 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/nfs.dump.dir in system properties and HBase conf 2023-08-10T06:31:55,438 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/java.io.tmpdir in system properties and HBase conf 2023-08-10T06:31:55,438 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:31:55,439 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2023-08-10T06:31:55,439 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/fs.s3a.committer.staging.tmp.path in system properties and HBase conf Formatting using clusterid: testClusterID 2023-08-10T06:31:55,887 WARN [Time-limited test {}] conf.Configuration(1733): No unit for dfs.heartbeat.interval(3) assuming SECONDS 2023-08-10T06:31:55,891 WARN [Time-limited test {}] conf.Configuration(1733): No unit for dfs.namenode.safemode.extension(0) assuming MILLISECONDS 2023-08-10T06:31:56,181 WARN [Time-limited test {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2023-08-10T06:31:56,347 INFO [Time-limited test {}] log.Slf4jLog(67): Logging to org.apache.logging.slf4j.Log4jLogger@18135890 via org.mortbay.log.Slf4jLog 2023-08-10T06:31:56,362 WARN [Time-limited test {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:31:56,400 INFO [Time-limited test {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:31:56,435 INFO [Time-limited test {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/hdfs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/java.io.tmpdir/Jetty_localhost_44251_hdfs____ijnlh4/webapp 2023-08-10T06:31:56,566 INFO [Time-limited test {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:44251 2023-08-10T06:31:56,588 WARN [Time-limited test {}] conf.Configuration(1733): No unit for dfs.heartbeat.interval(3) assuming SECONDS 2023-08-10T06:31:56,589 WARN [Time-limited test {}] conf.Configuration(1733): No unit for dfs.namenode.safemode.extension(0) assuming MILLISECONDS 2023-08-10T06:31:57,124 WARN [Listener at localhost/33549 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:31:57,191 WARN [Listener at localhost/33549 {}] conf.Configuration(1733): No unit for dfs.datanode.outliers.report.interval(1800000) assuming MILLISECONDS 2023-08-10T06:31:57,213 WARN [Listener at localhost/33549 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:31:57,221 INFO [Listener at localhost/33549 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:31:57,227 INFO [Listener at localhost/33549 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/datanode to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/java.io.tmpdir/Jetty_localhost_34201_datanode____1zw8t5/webapp 2023-08-10T06:31:57,331 INFO [Listener at localhost/33549 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:34201 2023-08-10T06:31:57,601 WARN [Listener at localhost/37579 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:31:57,884 WARN [Thread-52 {}] conf.Configuration(1733): No unit for dfs.datanode.lock-reporting-threshold-ms(300) assuming MILLISECONDS 2023-08-10T06:31:58,050 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x84e11773eca7bd71: Processing first storage report for DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0 from datanode 400c1f01-c3a4-43fd-9e9d-b52c8cc54f17 2023-08-10T06:31:58,052 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x84e11773eca7bd71: from storage DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0 node DatanodeRegistration(127.0.0.1:45413, datanodeUuid=400c1f01-c3a4-43fd-9e9d-b52c8cc54f17, infoPort=41755, infoSecurePort=0, ipcPort=37579, storageInfo=lv=-57;cid=testClusterID;nsid=1221029603;c=1691649115962), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2023-08-10T06:31:58,052 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x84e11773eca7bd71: Processing first storage report for DS-a5f8bcb0-53df-4367-b255-0a6de629eb78 from datanode 400c1f01-c3a4-43fd-9e9d-b52c8cc54f17 2023-08-10T06:31:58,052 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x84e11773eca7bd71: from storage DS-a5f8bcb0-53df-4367-b255-0a6de629eb78 node DatanodeRegistration(127.0.0.1:45413, datanodeUuid=400c1f01-c3a4-43fd-9e9d-b52c8cc54f17, infoPort=41755, infoSecurePort=0, ipcPort=37579, storageInfo=lv=-57;cid=testClusterID;nsid=1221029603;c=1691649115962), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:31:58,140 DEBUG [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9 2023-08-10T06:31:58,227 INFO [Listener at localhost/37579 {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/cluster_31ee86f8-f5ed-8a45-aa31-8b420ad4a168/zookeeper_0, clientPort=52413, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/cluster_31ee86f8-f5ed-8a45-aa31-8b420ad4a168/zookeeper_0/version-2, dataDirSize=424 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/cluster_31ee86f8-f5ed-8a45-aa31-8b420ad4a168/zookeeper_0/version-2, dataLogSize=424 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, serverId=0 2023-08-10T06:31:58,249 INFO [Listener at localhost/37579 {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=52413 2023-08-10T06:31:58,257 INFO [Listener at localhost/37579 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:31:58,259 INFO [Listener at localhost/37579 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:31:58,935 INFO [Listener at localhost/37579 {}] util.FSUtils(461): Created version file at hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 with version=8 2023-08-10T06:31:58,935 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(1460): Setting hbase.fs.tmp.dir to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging 2023-08-10T06:31:58,951 DEBUG [Listener at localhost/37579 {}] hbase.LocalHBaseCluster(134): Setting Master Port to random. 2023-08-10T06:31:58,951 DEBUG [Listener at localhost/37579 {}] hbase.LocalHBaseCluster(141): Setting RegionServer Port to random. 2023-08-10T06:31:58,952 DEBUG [Listener at localhost/37579 {}] hbase.LocalHBaseCluster(151): Setting RS InfoServer Port to random. 2023-08-10T06:31:58,952 DEBUG [Listener at localhost/37579 {}] hbase.LocalHBaseCluster(159): Setting Master InfoServer Port to random. 2023-08-10T06:31:59,160 DEBUG [Listener at localhost/37579 {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2023-08-10T06:31:59,314 INFO [Listener at localhost/37579 {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2023-08-10T06:31:59,955 INFO [Listener at localhost/37579 {}] client.ConnectionUtils(128): master/jenkins-hbase3:0 server-side Connection retries=45 2023-08-10T06:32:00,000 INFO [Listener at localhost/37579 {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:00,001 INFO [Listener at localhost/37579 {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:00,001 INFO [Listener at localhost/37579 {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2023-08-10T06:32:00,001 INFO [Listener at localhost/37579 {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:00,001 INFO [Listener at localhost/37579 {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2023-08-10T06:32:00,150 INFO [Listener at localhost/37579 {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService 2023-08-10T06:32:00,253 DEBUG [Listener at localhost/37579 {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2023-08-10T06:32:00,259 INFO [Listener at localhost/37579 {}] ipc.NettyRpcServer(177): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2023-08-10T06:32:00,307 DEBUG [Listener at localhost/37579 {}] channel.DefaultChannelId(79): -Dio.netty.processId: 21051 (auto-detected) 2023-08-10T06:32:00,308 DEBUG [Listener at localhost/37579 {}] channel.DefaultChannelId(101): -Dio.netty.machineId: 02:42:1c:ff:fe:e2:13:ea (auto-detected) 2023-08-10T06:32:00,347 INFO [Listener at localhost/37579 {}] ipc.NettyRpcServer(144): Bind to /172.31.12.81:42279 2023-08-10T06:32:00,366 INFO [Listener at localhost/37579 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:00,369 INFO [Listener at localhost/37579 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:00,387 INFO [Listener at localhost/37579 {}] zookeeper.RecoverableZooKeeper(142): Process identifier=master:42279 connecting to ZooKeeper ensemble=127.0.0.1:52413 2023-08-10T06:32:00,434 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:422790x0, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2023-08-10T06:32:00,438 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(630): master:42279-0x10060ee705b0000 connected 2023-08-10T06:32:00,468 DEBUG [Listener at localhost/37579 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/master 2023-08-10T06:32:00,472 DEBUG [Listener at localhost/37579 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/running 2023-08-10T06:32:00,475 DEBUG [Listener at localhost/37579 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/acl 2023-08-10T06:32:00,486 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=42279 2023-08-10T06:32:00,486 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=42279 2023-08-10T06:32:00,487 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=42279 2023-08-10T06:32:00,487 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=42279 2023-08-10T06:32:00,487 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=42279 2023-08-10T06:32:00,540 INFO [Listener at localhost/37579 {}] log.Log(170): Logging initialized @6797ms to org.apache.hbase.thirdparty.org.eclipse.jetty.util.log.Slf4jLog 2023-08-10T06:32:00,661 INFO [Listener at localhost/37579 {}] http.HttpServer(973): Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter) 2023-08-10T06:32:00,662 INFO [Listener at localhost/37579 {}] http.HttpServer(973): Added global filter 'clickjackingprevention' (class=org.apache.hadoop.hbase.http.ClickjackingPreventionFilter) 2023-08-10T06:32:00,663 INFO [Listener at localhost/37579 {}] http.HttpServer(973): Added global filter 'securityheaders' (class=org.apache.hadoop.hbase.http.SecurityHeadersFilter) 2023-08-10T06:32:00,665 INFO [Listener at localhost/37579 {}] http.HttpServer(952): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context master 2023-08-10T06:32:00,665 INFO [Listener at localhost/37579 {}] http.HttpServer(959): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2023-08-10T06:32:00,665 INFO [Listener at localhost/37579 {}] http.HttpServer(959): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2023-08-10T06:32:00,668 INFO [Listener at localhost/37579 {}] http.HttpServer(797): ASYNC_PROFILER_HOME environment variable and async.profiler.home system property not specified. Disabling /prof endpoint. 2023-08-10T06:32:00,798 INFO [Listener at localhost/37579 {}] http.HttpServer(1219): Jetty bound to port 40903 2023-08-10T06:32:00,800 INFO [Listener at localhost/37579 {}] server.Server(375): jetty-9.4.50.v20221201; built: 2022-12-01T22:07:03.915Z; git: da9a0b30691a45daf90a9f17b5defa2f1434f882; jvm 1.8.0_362-b09 2023-08-10T06:32:00,831 INFO [Listener at localhost/37579 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:00,836 INFO [Listener at localhost/37579 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.s.ServletContextHandler@4f27c993{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.log.dir/,AVAILABLE} 2023-08-10T06:32:00,837 INFO [Listener at localhost/37579 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:00,837 INFO [Listener at localhost/37579 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.s.ServletContextHandler@67965e92{static,/static,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/static/,AVAILABLE} 2023-08-10T06:32:00,903 INFO [Listener at localhost/37579 {}] webapp.StandardDescriptorProcessor(277): NO JSP Support for /, did not find org.apache.hbase.thirdparty.org.eclipse.jetty.jsp.JettyJspServlet 2023-08-10T06:32:00,917 INFO [Listener at localhost/37579 {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2023-08-10T06:32:00,917 INFO [Listener at localhost/37579 {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2023-08-10T06:32:00,919 INFO [Listener at localhost/37579 {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2023-08-10T06:32:00,927 INFO [Listener at localhost/37579 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:02,050 INFO [Listener at localhost/37579 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.w.WebAppContext@6bc05b4f{master,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/master/,AVAILABLE}{file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/master} 2023-08-10T06:32:02,061 INFO [Listener at localhost/37579 {}] server.AbstractConnector(333): Started ServerConnector@46005c78{HTTP/1.1, (http/1.1)}{0.0.0.0:40903} 2023-08-10T06:32:02,061 INFO [Listener at localhost/37579 {}] server.Server(415): Started @8318ms 2023-08-10T06:32:02,065 INFO [Listener at localhost/37579 {}] master.HMaster(484): hbase.rootdir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3, hbase.cluster.distributed=false 2023-08-10T06:32:02,120 INFO [Listener at localhost/37579 {}] client.ConnectionUtils(128): regionserver/jenkins-hbase3:0 server-side Connection retries=45 2023-08-10T06:32:02,120 INFO [Listener at localhost/37579 {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:02,120 INFO [Listener at localhost/37579 {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:02,120 INFO [Listener at localhost/37579 {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2023-08-10T06:32:02,120 INFO [Listener at localhost/37579 {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:02,121 INFO [Listener at localhost/37579 {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2023-08-10T06:32:02,128 INFO [Listener at localhost/37579 {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService 2023-08-10T06:32:02,130 INFO [Listener at localhost/37579 {}] ipc.NettyRpcServer(177): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2023-08-10T06:32:02,131 INFO [Listener at localhost/37579 {}] ipc.NettyRpcServer(144): Bind to /172.31.12.81:34991 2023-08-10T06:32:02,133 INFO [Listener at localhost/37579 {}] hfile.BlockCacheFactory(127): Allocating BlockCache size=782.40 MB, blockSize=64 KB 2023-08-10T06:32:02,140 DEBUG [Listener at localhost/37579 {}] mob.MobFileCache(120): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2023-08-10T06:32:02,141 INFO [Listener at localhost/37579 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:02,143 INFO [Listener at localhost/37579 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:02,145 INFO [Listener at localhost/37579 {}] zookeeper.RecoverableZooKeeper(142): Process identifier=regionserver:34991 connecting to ZooKeeper ensemble=127.0.0.1:52413 2023-08-10T06:32:02,148 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:349910x0, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2023-08-10T06:32:02,149 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(630): regionserver:34991-0x10060ee705b0001 connected 2023-08-10T06:32:02,149 DEBUG [Listener at localhost/37579 {}] zookeeper.ZKUtil(113): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/master 2023-08-10T06:32:02,150 DEBUG [Listener at localhost/37579 {}] zookeeper.ZKUtil(113): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/running 2023-08-10T06:32:02,151 DEBUG [Listener at localhost/37579 {}] zookeeper.ZKUtil(113): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/acl 2023-08-10T06:32:02,152 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34991 2023-08-10T06:32:02,152 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34991 2023-08-10T06:32:02,152 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34991 2023-08-10T06:32:02,153 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34991 2023-08-10T06:32:02,153 DEBUG [Listener at localhost/37579 {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34991 2023-08-10T06:32:02,156 INFO [Listener at localhost/37579 {}] http.HttpServer(973): Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter) 2023-08-10T06:32:02,156 INFO [Listener at localhost/37579 {}] http.HttpServer(973): Added global filter 'clickjackingprevention' (class=org.apache.hadoop.hbase.http.ClickjackingPreventionFilter) 2023-08-10T06:32:02,156 INFO [Listener at localhost/37579 {}] http.HttpServer(973): Added global filter 'securityheaders' (class=org.apache.hadoop.hbase.http.SecurityHeadersFilter) 2023-08-10T06:32:02,157 INFO [Listener at localhost/37579 {}] http.HttpServer(952): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context regionserver 2023-08-10T06:32:02,157 INFO [Listener at localhost/37579 {}] http.HttpServer(959): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2023-08-10T06:32:02,157 INFO [Listener at localhost/37579 {}] http.HttpServer(959): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2023-08-10T06:32:02,157 INFO [Listener at localhost/37579 {}] http.HttpServer(797): ASYNC_PROFILER_HOME environment variable and async.profiler.home system property not specified. Disabling /prof endpoint. 2023-08-10T06:32:02,159 INFO [Listener at localhost/37579 {}] http.HttpServer(1219): Jetty bound to port 33811 2023-08-10T06:32:02,160 INFO [Listener at localhost/37579 {}] server.Server(375): jetty-9.4.50.v20221201; built: 2022-12-01T22:07:03.915Z; git: da9a0b30691a45daf90a9f17b5defa2f1434f882; jvm 1.8.0_362-b09 2023-08-10T06:32:02,161 INFO [Listener at localhost/37579 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:02,161 INFO [Listener at localhost/37579 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.s.ServletContextHandler@3c8b8bc0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.log.dir/,AVAILABLE} 2023-08-10T06:32:02,162 INFO [Listener at localhost/37579 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:02,162 INFO [Listener at localhost/37579 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.s.ServletContextHandler@2b2d8bca{static,/static,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/static/,AVAILABLE} 2023-08-10T06:32:02,173 INFO [Listener at localhost/37579 {}] webapp.StandardDescriptorProcessor(277): NO JSP Support for /, did not find org.apache.hbase.thirdparty.org.eclipse.jetty.jsp.JettyJspServlet 2023-08-10T06:32:02,173 INFO [Listener at localhost/37579 {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2023-08-10T06:32:02,174 INFO [Listener at localhost/37579 {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2023-08-10T06:32:02,174 INFO [Listener at localhost/37579 {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2023-08-10T06:32:02,175 INFO [Listener at localhost/37579 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:02,179 INFO [Listener at localhost/37579 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.w.WebAppContext@4c765c3c{regionserver,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/regionserver/,AVAILABLE}{file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/regionserver} 2023-08-10T06:32:02,180 INFO [Listener at localhost/37579 {}] server.AbstractConnector(333): Started ServerConnector@5233fa39{HTTP/1.1, (http/1.1)}{0.0.0.0:33811} 2023-08-10T06:32:02,180 INFO [Listener at localhost/37579 {}] server.Server(415): Started @8437ms 2023-08-10T06:32:02,185 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] server.Server(375): jetty-9.4.50.v20221201; built: 2022-12-01T22:07:03.915Z; git: da9a0b30691a45daf90a9f17b5defa2f1434f882; jvm 1.8.0_362-b09 2023-08-10T06:32:02,188 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] server.AbstractConnector(333): Started ServerConnector@696643d9{HTTP/1.1, (http/1.1)}{0.0.0.0:42951} 2023-08-10T06:32:02,188 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] server.Server(415): Started @8445ms 2023-08-10T06:32:02,188 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(2387): Adding backup master ZNode /1/backup-masters/jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:02,200 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/backup-masters 2023-08-10T06:32:02,200 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/backup-masters 2023-08-10T06:32:02,202 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on existing znode=/1/backup-masters/jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:02,209 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;jenkins-hbase3:42279 2023-08-10T06:32:02,222 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/master 2023-08-10T06:32:02,222 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/master 2023-08-10T06:32:02,222 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:02,222 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:02,223 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on existing znode=/1/master 2023-08-10T06:32:02,224 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on existing znode=/1/master 2023-08-10T06:32:02,225 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /1/backup-masters/jenkins-hbase3.apache.org,42279,1691649119143 from backup master directory 2023-08-10T06:32:02,228 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/backup-masters/jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:02,228 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/backup-masters 2023-08-10T06:32:02,228 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/backup-masters 2023-08-10T06:32:02,229 WARN [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2023-08-10T06:32:02,229 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:02,232 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 352, initial count 0 2023-08-10T06:32:02,233 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 391, initial count 0 2023-08-10T06:32:02,311 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] util.FSUtils(610): Created cluster ID file at hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase.id with ID: bbab9112-1eff-4d3c-8041-5490dbe4c92d 2023-08-10T06:32:02,349 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:02,366 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:02,366 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:02,407 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegion(355): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:02,409 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2023-08-10T06:32:02,429 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(270): ClientProtocol::create wrong number of arguments, should be hadoop 3.2 or below 2023-08-10T06:32:02,429 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(276): ClientProtocol::create wrong number of arguments, should be hadoop 2.x 2023-08-10T06:32:02,430 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(285): can not find SHOULD_REPLICATE flag, should be hadoop 2.x java.lang.IllegalArgumentException: No enum constant org.apache.hadoop.fs.CreateFlag.SHOULD_REPLICATE at java.lang.Enum.valueOf(Enum.java:238) ~[?:1.8.0_362] at org.apache.hadoop.fs.CreateFlag.valueOf(CreateFlag.java:63) ~[hadoop-common-2.10.2.jar:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.loadShouldReplicateFlag(FanOutOneBlockAsyncDFSOutputHelper.java:283) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.(FanOutOneBlockAsyncDFSOutputHelper.java:310) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at java.lang.Class.forName0(Native Method) ~[?:1.8.0_362] at java.lang.Class.forName(Class.java:264) ~[?:1.8.0_362] at org.apache.hadoop.hbase.wal.AsyncFSWALProvider.load(AsyncFSWALProvider.java:140) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProviderClass(WALFactory.java:160) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProvider(WALFactory.java:200) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:232) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:207) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.create(MasterRegion.java:385) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegionFactory.create(MasterRegionFactory.java:135) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.finishActiveMasterInitialization(HMaster.java:950) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.startActiveMasterManager(HMaster.java:2412) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.lambda$null$0(HMaster.java:573) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) ~[hbase-common-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) ~[hbase-common-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.master.HMaster.lambda$run$1(HMaster.java:570) ~[classes/:?] at java.lang.Thread.run(Thread.java:750) ~[?:1.8.0_362] 2023-08-10T06:32:02,440 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2130) ~[?:1.8.0_362] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at java.lang.Class.forName0(Native Method) ~[?:1.8.0_362] at java.lang.Class.forName(Class.java:264) ~[?:1.8.0_362] at org.apache.hadoop.hbase.wal.AsyncFSWALProvider.load(AsyncFSWALProvider.java:141) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProviderClass(WALFactory.java:160) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProvider(WALFactory.java:200) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:232) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:207) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.create(MasterRegion.java:385) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegionFactory.create(MasterRegionFactory.java:135) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.finishActiveMasterInitialization(HMaster.java:950) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.startActiveMasterManager(HMaster.java:2412) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.lambda$null$0(HMaster.java:573) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) ~[hbase-common-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) ~[hbase-common-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.master.HMaster.lambda$run$1(HMaster.java:570) ~[classes/:?] at java.lang.Thread.run(Thread.java:750) ~[?:1.8.0_362] 2023-08-10T06:32:02,442 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2023-08-10T06:32:02,510 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(7093): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store 2023-08-10T06:32:02,933 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2023-08-10T06:32:02,934 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(888): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:02,934 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1668): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2023-08-10T06:32:02,934 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1690): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:32:02,935 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1711): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:32:02,935 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1778): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2023-08-10T06:32:02,935 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1788): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:32:02,935 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1909): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:32:02,935 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1622): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2023-08-10T06:32:02,937 WARN [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegion(232): failed to clean up initializing flag: hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/.initializing 2023-08-10T06:32:02,937 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegion(202): WALDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/WALs/jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:02,947 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2023-08-10T06:32:02,961 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.AbstractFSWAL(491): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=jenkins-hbase3.apache.org%2C42279%2C1691649119143, suffix=, logDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/WALs/jenkins-hbase3.apache.org,42279,1691649119143, archiveDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/oldWALs, maxLogs=10 2023-08-10T06:32:02,991 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/WALs/jenkins-hbase3.apache.org,42279,1691649119143/jenkins-hbase3.apache.org%2C42279%2C1691649119143.1691649122968, exclude list is [], retry=0 2023-08-10T06:32:03,017 DEBUG [RS-EventLoopGroup-3-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK] 2023-08-10T06:32:03,025 DEBUG [RS-EventLoopGroup-3-2 {}] asyncfs.ProtobufDecoder(123): Hadoop 3.2 and below use unshaded protobuf. java.lang.ClassNotFoundException: org.apache.hadoop.thirdparty.protobuf.MessageLite at java.net.URLClassLoader.findClass(URLClassLoader.java:387) ~[?:1.8.0_362] at java.lang.ClassLoader.loadClass(ClassLoader.java:418) ~[?:1.8.0_362] at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) ~[?:1.8.0_362] at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ~[?:1.8.0_362] at java.lang.Class.forName0(Native Method) ~[?:1.8.0_362] at java.lang.Class.forName(Class.java:264) ~[?:1.8.0_362] at org.apache.hadoop.hbase.io.asyncfs.ProtobufDecoder.(ProtobufDecoder.java:118) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.processWriteBlockResponse(FanOutOneBlockAsyncDFSOutputHelper.java:346) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.access$100(FanOutOneBlockAsyncDFSOutputHelper.java:120) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$4.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:430) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:557) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.addListener(DefaultPromise.java:185) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.addListener(DefaultPromise.java:35) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hadoop.hbase.util.NettyFutureUtils.addListener(NettyFutureUtils.java:52) ~[hbase-common-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:424) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.access$300(FanOutOneBlockAsyncDFSOutputHelper.java:120) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:482) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:477) ~[hbase-asyncfs-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:653) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:691) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:489) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:397) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.4.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.4.jar:?] at java.lang.Thread.run(Thread.java:750) ~[?:1.8.0_362] 2023-08-10T06:32:03,093 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.AbstractFSWAL(832): New WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/WALs/jenkins-hbase3.apache.org,42279,1691649119143/jenkins-hbase3.apache.org%2C42279%2C1691649119143.1691649122968 2023-08-10T06:32:03,093 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK]] 2023-08-10T06:32:03,094 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(7254): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:03,094 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(888): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:03,099 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(7296): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,100 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(7299): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,165 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,204 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2023-08-10T06:32:03,210 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:03,213 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(324): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:03,214 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,216 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2023-08-10T06:32:03,216 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:03,217 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(324): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:03,218 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,220 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2023-08-10T06:32:03,220 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:03,221 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(324): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:03,221 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,223 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2023-08-10T06:32:03,223 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:03,224 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(324): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:03,228 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,229 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,244 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2023-08-10T06:32:03,247 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1076): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:03,251 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:03,252 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1093): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=10953211360, jitterRate=0.02009730041027069}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2023-08-10T06:32:03,258 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(992): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2023-08-10T06:32:03,259 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2023-08-10T06:32:03,295 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@76db2cdb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:03,349 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(859): No meta location available on zookeeper, skip migrating... 2023-08-10T06:32:03,371 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2023-08-10T06:32:03,371 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(564): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2023-08-10T06:32:03,374 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2023-08-10T06:32:03,376 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(584): Recovered RegionProcedureStore lease in 1 msec 2023-08-10T06:32:03,382 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(598): Loaded RegionProcedureStore in 6 msec 2023-08-10T06:32:03,382 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2023-08-10T06:32:03,415 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2023-08-10T06:32:03,431 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Unable to get data of znode /1/balancer because node does not exist (not necessarily an error) 2023-08-10T06:32:03,433 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /1/balancer already deleted, retry=false 2023-08-10T06:32:03,436 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(150): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2023-08-10T06:32:03,437 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Unable to get data of znode /1/normalizer because node does not exist (not necessarily an error) 2023-08-10T06:32:03,438 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /1/normalizer already deleted, retry=false 2023-08-10T06:32:03,441 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2023-08-10T06:32:03,447 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Unable to get data of znode /1/switch/split because node does not exist (not necessarily an error) 2023-08-10T06:32:03,449 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /1/switch/split already deleted, retry=false 2023-08-10T06:32:03,450 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Unable to get data of znode /1/switch/merge because node does not exist (not necessarily an error) 2023-08-10T06:32:03,452 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /1/switch/merge already deleted, retry=false 2023-08-10T06:32:03,467 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Unable to get data of znode /1/snapshot-cleanup because node does not exist (not necessarily an error) 2023-08-10T06:32:03,468 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /1/snapshot-cleanup already deleted, retry=false 2023-08-10T06:32:03,474 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/running 2023-08-10T06:32:03,474 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/running 2023-08-10T06:32:03,474 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:03,474 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:03,474 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(803): Active/primary master=jenkins-hbase3.apache.org,42279,1691649119143, sessionid=0x10060ee705b0000, setting cluster-up flag (Was=false) 2023-08-10T06:32:03,483 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ProcedureManagerHost(66): User procedure org.apache.hadoop.hbase.backup.master.LogRollMasterProcedureManager was loaded successfully. 2023-08-10T06:32:03,491 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:03,491 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:03,497 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /1/flush-table-proc/acquired, /1/flush-table-proc/reached, /1/flush-table-proc/abort 2023-08-10T06:32:03,499 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:03,503 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:03,503 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:03,510 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /1/rolllog-proc/acquired, /1/rolllog-proc/reached, /1/rolllog-proc/abort 2023-08-10T06:32:03,511 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:03,517 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:03,518 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:03,525 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /1/online-snapshot/acquired, /1/online-snapshot/reached, /1/online-snapshot/abort 2023-08-10T06:32:03,527 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:03,594 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;jenkins-hbase3:34991 2023-08-10T06:32:03,595 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(997): ClusterId : bbab9112-1eff-4d3c-8041-5490dbe4c92d 2023-08-10T06:32:03,597 INFO [RS:0;jenkins-hbase3:34991 {}] procedure.ProcedureManagerHost(66): User procedure org.apache.hadoop.hbase.backup.regionserver.LogRollRegionServerProcedureManager was loaded successfully. 2023-08-10T06:32:03,599 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2023-08-10T06:32:03,605 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2023-08-10T06:32:03,606 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(43): Procedure backup-proc initializing 2023-08-10T06:32:03,610 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(45): Procedure backup-proc initialized 2023-08-10T06:32:03,610 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2023-08-10T06:32:03,614 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2023-08-10T06:32:03,615 DEBUG [RS:0;jenkins-hbase3:34991 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4283bbfe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:03,617 DEBUG [RS:0;jenkins-hbase3:34991 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@73f6d30b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=jenkins-hbase3.apache.org/172.31.12.81:0 2023-08-10T06:32:03,624 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2023-08-10T06:32:03,624 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2023-08-10T06:32:03,624 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1079): About to register with Master. 2023-08-10T06:32:03,628 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3023): reportForDuty to master=jenkins-hbase3.apache.org,42279,1691649119143 with isa=jenkins-hbase3.apache.org/172.31.12.81:34991, startcode=1691649122119 2023-08-10T06:32:03,647 DEBUG [RS:0;jenkins-hbase3:34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2023-08-10T06:32:03,680 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1033): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2023-08-10T06:32:03,689 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(574): slop=0.2 2023-08-10T06:32:03,696 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(285): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2023-08-10T06:32:03,707 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: jenkins-hbase3.apache.org,42279,1691649119143 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2023-08-10T06:32:03,714 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/jenkins-hbase3:0, corePoolSize=5, maxPoolSize=5 2023-08-10T06:32:03,714 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/jenkins-hbase3:0, corePoolSize=5, maxPoolSize=5 2023-08-10T06:32:03,714 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=5, maxPoolSize=5 2023-08-10T06:32:03,715 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=5, maxPoolSize=5 2023-08-10T06:32:03,715 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/jenkins-hbase3:0, corePoolSize=10, maxPoolSize=10 2023-08-10T06:32:03,716 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,716 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=2, maxPoolSize=2 2023-08-10T06:32:03,716 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,723 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1691649153722 2023-08-10T06:32:03,727 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:45161, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2023-08-10T06:32:03,727 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2023-08-10T06:32:03,732 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2023-08-10T06:32:03,732 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2023-08-10T06:32:03,733 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2023-08-10T06:32:03,737 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42279 {}] ipc.MetricsHBaseServer(144): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3180) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:588) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:17393) ~[hbase-protocol-shaded-2.6.0-SNAPSHOT.jar:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:437) ~[classes/:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.6.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.6.0-SNAPSHOT] 2023-08-10T06:32:03,739 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:03,739 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2023-08-10T06:32:03,746 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2023-08-10T06:32:03,747 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.backup.master.BackupLogCleaner 2023-08-10T06:32:03,749 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2023-08-10T06:32:03,751 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2023-08-10T06:32:03,751 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2023-08-10T06:32:03,758 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,759 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2023-08-10T06:32:03,761 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2023-08-10T06:32:03,761 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2023-08-10T06:32:03,768 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ReadOnlyZKClient(139): Connect 0x68964ef7 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:03,773 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3042): Master is not running yet 2023-08-10T06:32:03,774 WARN [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1088): reportForDuty failed; sleeping 100 ms and then retrying. 2023-08-10T06:32:03,778 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1686575d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:03,778 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.backup.BackupHFileCleaner 2023-08-10T06:32:03,782 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2023-08-10T06:32:03,782 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2023-08-10T06:32:03,785 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.large.0-1691649123785,5,FailOnTimeoutGroup] 2023-08-10T06:32:03,785 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.small.0-1691649123785,5,FailOnTimeoutGroup] 2023-08-10T06:32:03,785 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,785 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(1634): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2023-08-10T06:32:03,787 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,787 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,875 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3023): reportForDuty to master=jenkins-hbase3.apache.org,42279,1691649119143 with isa=jenkins-hbase3.apache.org/172.31.12.81:34991, startcode=1691649122119 2023-08-10T06:32:03,881 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42279 {}] master.ServerManager(403): Registering regionserver=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:03,894 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1705): Config from master: hbase.rootdir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:03,894 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1705): Config from master: fs.defaultFS=hdfs://localhost:33549 2023-08-10T06:32:03,895 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1705): Config from master: hbase.master.info.port=40903 2023-08-10T06:32:03,902 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rs 2023-08-10T06:32:03,902 DEBUG [RS:0;jenkins-hbase3:34991 {}] zookeeper.ZKUtil(111): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on existing znode=/1/rs/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:03,902 WARN [RS:0;jenkins-hbase3:34991 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2023-08-10T06:32:03,903 INFO [RS:0;jenkins-hbase3:34991 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2023-08-10T06:32:03,903 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(2060): logDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:03,905 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [jenkins-hbase3.apache.org,34991,1691649122119] 2023-08-10T06:32:03,927 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2023-08-10T06:32:03,940 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.MetricsRegionServerWrapperImpl(165): Computing regionserver metrics every 5000 milliseconds 2023-08-10T06:32:03,961 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=782.4 M, globalMemStoreLimitLowMark=743.3 M, Offheap=false 2023-08-10T06:32:03,967 INFO [RS:0;jenkins-hbase3:34991 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2023-08-10T06:32:03,967 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,970 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer$CompactionChecker(1947): CompactionChecker runs every PT1S 2023-08-10T06:32:03,982 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,982 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,983 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,983 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,983 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,983 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,984 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/jenkins-hbase3:0, corePoolSize=2, maxPoolSize=2 2023-08-10T06:32:03,984 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,984 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,984 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,985 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,985 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:03,985 DEBUG [RS:0;jenkins-hbase3:34991 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0, corePoolSize=3, maxPoolSize=3 2023-08-10T06:32:03,988 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,988 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,988 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,988 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:03,988 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,34991,1691649122119-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2023-08-10T06:32:04,006 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2023-08-10T06:32:04,009 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,34991,1691649122119-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:04,030 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.Replication(204): jenkins-hbase3.apache.org,34991,1691649122119 started 2023-08-10T06:32:04,030 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1747): Serving as jenkins-hbase3.apache.org,34991,1691649122119, RpcServer on jenkins-hbase3.apache.org/172.31.12.81:34991, sessionid=0x10060ee705b0001 2023-08-10T06:32:04,030 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2023-08-10T06:32:04,031 DEBUG [RS:0;jenkins-hbase3:34991 {}] flush.RegionServerFlushTableProcedureManager(106): Start region server flush procedure manager jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:04,031 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'jenkins-hbase3.apache.org,34991,1691649122119' 2023-08-10T06:32:04,031 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/1/flush-table-proc/abort' 2023-08-10T06:32:04,031 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/1/flush-table-proc/acquired' 2023-08-10T06:32:04,032 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2023-08-10T06:32:04,032 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(51): Procedure backup-proc starting 2023-08-10T06:32:04,032 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'jenkins-hbase3.apache.org,34991,1691649122119' 2023-08-10T06:32:04,032 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/1/rolllog-proc/abort' 2023-08-10T06:32:04,033 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/1/rolllog-proc/acquired' 2023-08-10T06:32:04,033 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.LogRollRegionServerProcedureManager(92): Started region server backup manager. 2023-08-10T06:32:04,033 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(53): Procedure backup-proc started 2023-08-10T06:32:04,033 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2023-08-10T06:32:04,033 DEBUG [RS:0;jenkins-hbase3:34991 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:04,033 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'jenkins-hbase3.apache.org,34991,1691649122119' 2023-08-10T06:32:04,033 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/1/online-snapshot/abort' 2023-08-10T06:32:04,034 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/1/online-snapshot/acquired' 2023-08-10T06:32:04,034 DEBUG [RS:0;jenkins-hbase3:34991 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2023-08-10T06:32:04,035 INFO [RS:0;jenkins-hbase3:34991 {}] quotas.RegionServerRpcQuotaManager(63): Quota support disabled 2023-08-10T06:32:04,035 INFO [RS:0;jenkins-hbase3:34991 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2023-08-10T06:32:04,145 INFO [RS:0;jenkins-hbase3:34991 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2023-08-10T06:32:04,148 INFO [RS:0;jenkins-hbase3:34991 {}] wal.AbstractFSWAL(491): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=jenkins-hbase3.apache.org%2C34991%2C1691649122119, suffix=, logDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119, archiveDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/oldWALs, maxLogs=32 2023-08-10T06:32:04,156 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2023-08-10T06:32:04,156 INFO [PEWorker-1 {}] regionserver.HRegion(7075): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:04,161 DEBUG [RS:0;jenkins-hbase3:34991 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649124149, exclude list is [], retry=0 2023-08-10T06:32:04,170 DEBUG [RS-EventLoopGroup-3-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK] 2023-08-10T06:32:04,174 INFO [RS:0;jenkins-hbase3:34991 {}] wal.AbstractFSWAL(832): New WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649124149 2023-08-10T06:32:04,174 DEBUG [RS:0;jenkins-hbase3:34991 {}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK]] 2023-08-10T06:32:04,573 DEBUG [PEWorker-1 {}] regionserver.HRegion(888): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:04,575 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2023-08-10T06:32:04,578 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2023-08-10T06:32:04,578 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:04,579 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:04,579 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2023-08-10T06:32:04,581 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2023-08-10T06:32:04,581 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:04,582 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:04,582 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2023-08-10T06:32:04,584 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2023-08-10T06:32:04,584 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:04,585 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:04,586 DEBUG [PEWorker-1 {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740 2023-08-10T06:32:04,587 DEBUG [PEWorker-1 {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740 2023-08-10T06:32:04,589 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2023-08-10T06:32:04,591 DEBUG [PEWorker-1 {}] regionserver.HRegion(1076): writing seq id for 1588230740 2023-08-10T06:32:04,594 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:04,595 INFO [PEWorker-1 {}] regionserver.HRegion(1093): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11398557120, jitterRate=0.06157335638999939}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2023-08-10T06:32:04,597 DEBUG [PEWorker-1 {}] regionserver.HRegion(992): Region open journal for 1588230740: 2023-08-10T06:32:04,597 DEBUG [PEWorker-1 {}] regionserver.HRegion(1668): Closing 1588230740, disabling compactions & flushes 2023-08-10T06:32:04,597 INFO [PEWorker-1 {}] regionserver.HRegion(1690): Closing region hbase:meta,,1.1588230740 2023-08-10T06:32:04,597 DEBUG [PEWorker-1 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2023-08-10T06:32:04,597 DEBUG [PEWorker-1 {}] regionserver.HRegion(1778): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2023-08-10T06:32:04,597 DEBUG [PEWorker-1 {}] regionserver.HRegion(1788): Updates disabled for region hbase:meta,,1.1588230740 2023-08-10T06:32:04,598 INFO [PEWorker-1 {}] regionserver.HRegion(1909): Closed hbase:meta,,1.1588230740 2023-08-10T06:32:04,599 DEBUG [PEWorker-1 {}] regionserver.HRegion(1622): Region close journal for 1588230740: 2023-08-10T06:32:04,601 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2023-08-10T06:32:04,602 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2023-08-10T06:32:04,610 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2023-08-10T06:32:04,623 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2023-08-10T06:32:04,628 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(262): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2023-08-10T06:32:04,780 DEBUG [jenkins-hbase3:42279 {}] assignment.AssignmentManager(2303): Processing assignQueue; systemServersCount=1, allServersCount=1 2023-08-10T06:32:04,788 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:04,794 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as jenkins-hbase3.apache.org,34991,1691649122119, state=OPENING 2023-08-10T06:32:04,805 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2023-08-10T06:32:04,811 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:04,811 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:04,811 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /1/meta-region-server: CHANGED 2023-08-10T06:32:04,811 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /1/meta-region-server: CHANGED 2023-08-10T06:32:04,815 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:05,006 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(721): New admin connection to jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:05,010 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=AdminService, sasl=false 2023-08-10T06:32:05,033 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:52982, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2023-08-10T06:32:05,048 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(130): Open hbase:meta,,1.1588230740 2023-08-10T06:32:05,049 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2023-08-10T06:32:05,049 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2023-08-10T06:32:05,052 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(491): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta, suffix=.meta, logDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119, archiveDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/oldWALs, maxLogs=32 2023-08-10T06:32:05,064 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649125053.meta, exclude list is [], retry=0 2023-08-10T06:32:05,070 DEBUG [RS-EventLoopGroup-3-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK] 2023-08-10T06:32:05,073 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(832): New WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649125053.meta 2023-08-10T06:32:05,073 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK]] 2023-08-10T06:32:05,074 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7254): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:05,079 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:05,079 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2023-08-10T06:32:05,096 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7955): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2023-08-10T06:32:05,096 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(393): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2023-08-10T06:32:05,101 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2023-08-10T06:32:05,101 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(888): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:05,102 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7296): checking encryption for 1588230740 2023-08-10T06:32:05,102 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7299): checking classloading for 1588230740 2023-08-10T06:32:05,104 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2023-08-10T06:32:05,106 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2023-08-10T06:32:05,106 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:05,107 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:05,107 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2023-08-10T06:32:05,108 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2023-08-10T06:32:05,108 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:05,109 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:05,109 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2023-08-10T06:32:05,110 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2023-08-10T06:32:05,111 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:05,111 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:05,113 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740 2023-08-10T06:32:05,115 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740 2023-08-10T06:32:05,118 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2023-08-10T06:32:05,120 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1076): writing seq id for 1588230740 2023-08-10T06:32:05,121 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11642225760, jitterRate=0.0842667669057846}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2023-08-10T06:32:05,123 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(992): Region open journal for 1588230740: 2023-08-10T06:32:05,134 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2552): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1691649124998 2023-08-10T06:32:05,155 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2579): Finished post open deploy task for hbase:meta,,1.1588230740 2023-08-10T06:32:05,156 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(158): Opened hbase:meta,,1.1588230740 2023-08-10T06:32:05,157 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:05,159 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as jenkins-hbase3.apache.org,34991,1691649122119, state=OPEN 2023-08-10T06:32:05,162 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/1/meta-region-server 2023-08-10T06:32:05,162 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/1/meta-region-server 2023-08-10T06:32:05,162 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /1/meta-region-server: CHANGED 2023-08-10T06:32:05,162 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /1/meta-region-server: CHANGED 2023-08-10T06:32:05,166 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=3, resume processing ppid=2 2023-08-10T06:32:05,166 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=jenkins-hbase3.apache.org,34991,1691649122119 in 347 msec 2023-08-10T06:32:05,175 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=2, resume processing ppid=1 2023-08-10T06:32:05,175 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 557 msec 2023-08-10T06:32:05,180 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 1.5810 sec 2023-08-10T06:32:05,180 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(1062): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1691649125180, completionTime=-1 2023-08-10T06:32:05,180 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.ServerManager(830): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2023-08-10T06:32:05,180 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] assignment.AssignmentManager(1629): Joining cluster... 2023-08-10T06:32:05,371 DEBUG [hconnection-0x21d01e4e-shared-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:05,376 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:52988, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:05,400 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] assignment.AssignmentManager(1641): Number of RegionServers=1 2023-08-10T06:32:05,400 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1691649185400 2023-08-10T06:32:05,400 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1691649245400 2023-08-10T06:32:05,401 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] assignment.AssignmentManager(1648): Joined the cluster in 220 msec 2023-08-10T06:32:05,436 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,42279,1691649119143-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:05,437 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,42279,1691649119143-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:05,437 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,42279,1691649119143-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:05,440 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-jenkins-hbase3:42279, period=300000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:05,440 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:05,448 DEBUG [master/jenkins-hbase3:0.Chore.1 {}] janitor.CatalogJanitor(175): 2023-08-10T06:32:05,463 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2023-08-10T06:32:05,465 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(2367): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2023-08-10T06:32:05,475 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1033): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2023-08-10T06:32:05,479 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:05,481 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:05,483 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:05,903 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => 333dde743887b960947ea82e4c5cfa5e, NAME => 'hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:06,316 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(888): Instantiated hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:06,317 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1668): Closing 333dde743887b960947ea82e4c5cfa5e, disabling compactions & flushes 2023-08-10T06:32:06,317 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1690): Closing region hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:32:06,317 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:32:06,317 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. after waiting 0 ms 2023-08-10T06:32:06,317 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:32:06,317 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1909): Closed hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:32:06,317 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1622): Region close journal for 333dde743887b960947ea82e4c5cfa5e: 2023-08-10T06:32:06,319 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:06,335 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1691649126322"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649126322"}]},"ts":"1691649126322"} 2023-08-10T06:32:06,376 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:06,379 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:06,385 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649126379"}]},"ts":"1691649126379"} 2023-08-10T06:32:06,389 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2023-08-10T06:32:06,395 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=333dde743887b960947ea82e4c5cfa5e, ASSIGN}] 2023-08-10T06:32:06,398 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=333dde743887b960947ea82e4c5cfa5e, ASSIGN 2023-08-10T06:32:06,399 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(262): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=333dde743887b960947ea82e4c5cfa5e, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:06,550 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=333dde743887b960947ea82e4c5cfa5e, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:06,554 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 333dde743887b960947ea82e4c5cfa5e, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:06,713 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(130): Open hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:32:06,713 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7254): Opening region: {ENCODED => 333dde743887b960947ea82e4c5cfa5e, NAME => 'hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:06,714 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:06,715 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:32:06,715 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(888): Instantiated hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:06,715 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7296): checking encryption for 333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:32:06,715 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7299): checking classloading for 333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:32:06,717 INFO [StoreOpener-333dde743887b960947ea82e4c5cfa5e-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:32:06,719 INFO [StoreOpener-333dde743887b960947ea82e4c5cfa5e-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 333dde743887b960947ea82e4c5cfa5e columnFamilyName info 2023-08-10T06:32:06,719 DEBUG [StoreOpener-333dde743887b960947ea82e4c5cfa5e-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:06,720 INFO [StoreOpener-333dde743887b960947ea82e4c5cfa5e-1 {}] regionserver.HStore(324): Store=333dde743887b960947ea82e4c5cfa5e/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:06,722 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:32:06,722 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:32:06,726 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1076): writing seq id for 333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:32:06,729 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:06,730 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1093): Opened 333dde743887b960947ea82e4c5cfa5e; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11163360960, jitterRate=0.03966900706291199}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2023-08-10T06:32:06,731 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(992): Region open journal for 333dde743887b960947ea82e4c5cfa5e: 2023-08-10T06:32:06,733 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2552): Post open deploy tasks for hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e., pid=6, masterSystemTime=1691649126707 2023-08-10T06:32:06,736 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2579): Finished post open deploy task for hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:32:06,736 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(158): Opened hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:32:06,737 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=333dde743887b960947ea82e4c5cfa5e, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:06,744 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=6, resume processing ppid=5 2023-08-10T06:32:06,744 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 333dde743887b960947ea82e4c5cfa5e, server=jenkins-hbase3.apache.org,34991,1691649122119 in 186 msec 2023-08-10T06:32:06,747 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=5, resume processing ppid=4 2023-08-10T06:32:06,747 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1414): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=333dde743887b960947ea82e4c5cfa5e, ASSIGN in 349 msec 2023-08-10T06:32:06,748 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:06,748 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649126748"}]},"ts":"1691649126748"} 2023-08-10T06:32:06,751 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2023-08-10T06:32:06,754 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:06,757 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 1.2880 sec 2023-08-10T06:32:06,780 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/namespace 2023-08-10T06:32:06,781 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/namespace 2023-08-10T06:32:06,781 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:06,781 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:32:06,837 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1033): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2023-08-10T06:32:06,854 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/namespace 2023-08-10T06:32:06,859 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 29 msec 2023-08-10T06:32:06,862 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1033): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2023-08-10T06:32:06,873 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/namespace 2023-08-10T06:32:06,878 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 15 msec 2023-08-10T06:32:06,888 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/1/namespace/default 2023-08-10T06:32:06,891 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/1/namespace/hbase 2023-08-10T06:32:06,894 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(1192): Master has completed initialization 4.665sec 2023-08-10T06:32:06,896 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2023-08-10T06:32:06,897 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2023-08-10T06:32:06,899 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2023-08-10T06:32:06,899 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2023-08-10T06:32:06,899 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2023-08-10T06:32:06,901 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,42279,1691649119143-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2023-08-10T06:32:06,902 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,42279,1691649119143-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2023-08-10T06:32:06,911 DEBUG [Listener at localhost/37579 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x2e3dc5ff to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:06,915 DEBUG [Listener at localhost/37579 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@875f62a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:06,920 DEBUG [Listener at localhost/37579 {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2023-08-10T06:32:06,921 DEBUG [Listener at localhost/37579 {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2023-08-10T06:32:06,921 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(1295): Balancer post startup initialization complete, took 0 seconds 2023-08-10T06:32:06,923 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2023-08-10T06:32:06,950 DEBUG [hconnection-0x32380546-shared-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:06,962 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:52996, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:06,974 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:32:06,986 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.log.dir so I do NOT create it in target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec 2023-08-10T06:32:06,986 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.tmp.dir so I do NOT create it in target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec 2023-08-10T06:32:06,986 DEBUG [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(348): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec 2023-08-10T06:32:06,986 INFO [Listener at localhost/37579 {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/cluster_8430befc-bdf7-6f28-aaa6-942b3e94ea59, deleteOnExit=true 2023-08-10T06:32:06,986 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/test.cache.data in system properties and HBase conf 2023-08-10T06:32:06,987 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop.tmp.dir in system properties and HBase conf 2023-08-10T06:32:06,987 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop.log.dir in system properties and HBase conf 2023-08-10T06:32:06,987 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/mapreduce.cluster.local.dir in system properties and HBase conf 2023-08-10T06:32:06,987 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/mapreduce.cluster.temp.dir in system properties and HBase conf 2023-08-10T06:32:06,987 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2023-08-10T06:32:06,987 DEBUG [Listener at localhost/37579 {}] fs.HFileSystem(308): The file system is not a DistributedFileSystem. Skipping on block location reordering 2023-08-10T06:32:06,987 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:32:06,988 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:32:06,988 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/yarn.nodemanager.log-dirs in system properties and HBase conf 2023-08-10T06:32:06,988 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:32:06,988 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2023-08-10T06:32:06,988 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2023-08-10T06:32:06,988 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:32:06,988 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:32:06,988 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2023-08-10T06:32:06,989 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/nfs.dump.dir in system properties and HBase conf 2023-08-10T06:32:06,989 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/java.io.tmpdir in system properties and HBase conf 2023-08-10T06:32:06,989 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:32:06,989 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2023-08-10T06:32:06,989 INFO [Listener at localhost/37579 {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/fs.s3a.committer.staging.tmp.path in system properties and HBase conf Formatting using clusterid: testClusterID 2023-08-10T06:32:07,002 WARN [Listener at localhost/37579 {}] conf.Configuration(1733): No unit for dfs.heartbeat.interval(3) assuming SECONDS 2023-08-10T06:32:07,002 WARN [Listener at localhost/37579 {}] conf.Configuration(1733): No unit for dfs.namenode.safemode.extension(0) assuming MILLISECONDS 2023-08-10T06:32:07,061 WARN [Listener at localhost/37579 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:07,065 INFO [Listener at localhost/37579 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:07,072 INFO [Listener at localhost/37579 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/hdfs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/java.io.tmpdir/Jetty_localhost_35369_hdfs____.xh7y6a/webapp 2023-08-10T06:32:07,174 INFO [Listener at localhost/37579 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:35369 2023-08-10T06:32:07,188 WARN [Listener at localhost/37579 {}] conf.Configuration(1733): No unit for dfs.heartbeat.interval(3) assuming SECONDS 2023-08-10T06:32:07,188 WARN [Listener at localhost/37579 {}] conf.Configuration(1733): No unit for dfs.namenode.safemode.extension(0) assuming MILLISECONDS 2023-08-10T06:32:07,235 WARN [Listener at localhost/44455 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:32:07,257 WARN [Listener at localhost/44455 {}] conf.Configuration(1733): No unit for dfs.datanode.outliers.report.interval(1800000) assuming MILLISECONDS 2023-08-10T06:32:07,260 WARN [Listener at localhost/44455 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:07,261 INFO [Listener at localhost/44455 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:07,268 INFO [Listener at localhost/44455 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/datanode to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/java.io.tmpdir/Jetty_localhost_40665_datanode____h9utpc/webapp 2023-08-10T06:32:07,366 INFO [Listener at localhost/44455 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:40665 2023-08-10T06:32:07,378 WARN [Listener at localhost/40993 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:32:07,423 WARN [Listener at localhost/40993 {}] conf.Configuration(1733): No unit for dfs.datanode.outliers.report.interval(1800000) assuming MILLISECONDS 2023-08-10T06:32:07,425 WARN [Listener at localhost/40993 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:07,427 INFO [Listener at localhost/40993 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:07,433 INFO [Listener at localhost/40993 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/datanode to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/java.io.tmpdir/Jetty_localhost_46745_datanode____89nj19/webapp 2023-08-10T06:32:07,489 WARN [Thread-202 {}] conf.Configuration(1733): No unit for dfs.datanode.lock-reporting-threshold-ms(300) assuming MILLISECONDS 2023-08-10T06:32:07,534 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x6c8c7533c10d091a: Processing first storage report for DS-3ea9105e-de54-44d1-a029-d02e873daf8c from datanode 294aae14-1491-4c05-bc43-e02619891c37 2023-08-10T06:32:07,534 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x6c8c7533c10d091a: from storage DS-3ea9105e-de54-44d1-a029-d02e873daf8c node DatanodeRegistration(127.0.0.1:34143, datanodeUuid=294aae14-1491-4c05-bc43-e02619891c37, infoPort=36621, infoSecurePort=0, ipcPort=40993, storageInfo=lv=-57;cid=testClusterID;nsid=1014535546;c=1691649127005), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2023-08-10T06:32:07,534 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x6c8c7533c10d091a: Processing first storage report for DS-460bef7d-e8a7-4384-9630-e7a4b7042f39 from datanode 294aae14-1491-4c05-bc43-e02619891c37 2023-08-10T06:32:07,534 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x6c8c7533c10d091a: from storage DS-460bef7d-e8a7-4384-9630-e7a4b7042f39 node DatanodeRegistration(127.0.0.1:34143, datanodeUuid=294aae14-1491-4c05-bc43-e02619891c37, infoPort=36621, infoSecurePort=0, ipcPort=40993, storageInfo=lv=-57;cid=testClusterID;nsid=1014535546;c=1691649127005), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:07,558 INFO [Listener at localhost/40993 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:46745 2023-08-10T06:32:07,568 WARN [Listener at localhost/42335 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:32:07,589 WARN [Listener at localhost/42335 {}] conf.Configuration(1733): No unit for dfs.datanode.outliers.report.interval(1800000) assuming MILLISECONDS 2023-08-10T06:32:07,591 WARN [Listener at localhost/42335 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:07,592 INFO [Listener at localhost/42335 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:07,598 INFO [Listener at localhost/42335 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/datanode to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/java.io.tmpdir/Jetty_localhost_38389_datanode____tie6hi/webapp 2023-08-10T06:32:07,643 WARN [Thread-237 {}] conf.Configuration(1733): No unit for dfs.datanode.lock-reporting-threshold-ms(300) assuming MILLISECONDS 2023-08-10T06:32:07,677 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x9262da7c5515ed83: Processing first storage report for DS-9386f1a7-3604-4ae3-9aed-a633fa75b158 from datanode 868b385d-1a59-4611-8f2d-3e9d2b8e4e1e 2023-08-10T06:32:07,677 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x9262da7c5515ed83: from storage DS-9386f1a7-3604-4ae3-9aed-a633fa75b158 node DatanodeRegistration(127.0.0.1:42505, datanodeUuid=868b385d-1a59-4611-8f2d-3e9d2b8e4e1e, infoPort=42647, infoSecurePort=0, ipcPort=42335, storageInfo=lv=-57;cid=testClusterID;nsid=1014535546;c=1691649127005), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:07,677 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x9262da7c5515ed83: Processing first storage report for DS-bf6a086e-932c-4218-98ab-2aa630e2f4b1 from datanode 868b385d-1a59-4611-8f2d-3e9d2b8e4e1e 2023-08-10T06:32:07,677 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x9262da7c5515ed83: from storage DS-bf6a086e-932c-4218-98ab-2aa630e2f4b1 node DatanodeRegistration(127.0.0.1:42505, datanodeUuid=868b385d-1a59-4611-8f2d-3e9d2b8e4e1e, infoPort=42647, infoSecurePort=0, ipcPort=42335, storageInfo=lv=-57;cid=testClusterID;nsid=1014535546;c=1691649127005), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:07,711 INFO [Listener at localhost/42335 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:38389 2023-08-10T06:32:07,723 WARN [Listener at localhost/36659 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:32:07,798 WARN [Thread-271 {}] conf.Configuration(1733): No unit for dfs.datanode.lock-reporting-threshold-ms(300) assuming MILLISECONDS 2023-08-10T06:32:07,823 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x2afa42f58f9b4d27: Processing first storage report for DS-660e422a-4614-4be2-bdae-34ba7271ed00 from datanode b86086af-0069-40ac-b54c-a5ce0bf8379b 2023-08-10T06:32:07,823 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x2afa42f58f9b4d27: from storage DS-660e422a-4614-4be2-bdae-34ba7271ed00 node DatanodeRegistration(127.0.0.1:37669, datanodeUuid=b86086af-0069-40ac-b54c-a5ce0bf8379b, infoPort=38811, infoSecurePort=0, ipcPort=36659, storageInfo=lv=-57;cid=testClusterID;nsid=1014535546;c=1691649127005), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:07,823 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x2afa42f58f9b4d27: Processing first storage report for DS-81496f65-da76-41ea-98d9-69b40e6686ce from datanode b86086af-0069-40ac-b54c-a5ce0bf8379b 2023-08-10T06:32:07,823 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x2afa42f58f9b4d27: from storage DS-81496f65-da76-41ea-98d9-69b40e6686ce node DatanodeRegistration(127.0.0.1:37669, datanodeUuid=b86086af-0069-40ac-b54c-a5ce0bf8379b, infoPort=38811, infoSecurePort=0, ipcPort=36659, storageInfo=lv=-57;cid=testClusterID;nsid=1014535546;c=1691649127005), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:07,833 DEBUG [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec 2023-08-10T06:32:07,833 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=1, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2023-08-10T06:32:07,833 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(1143): NOT STARTING DFS 2023-08-10T06:32:07,834 INFO [Listener at localhost/36659 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:07,835 INFO [Listener at localhost/36659 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:07,856 INFO [Listener at localhost/36659 {}] util.FSUtils(461): Created version file at hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa with version=8 2023-08-10T06:32:07,856 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging 2023-08-10T06:32:07,857 DEBUG [Listener at localhost/36659 {}] hbase.LocalHBaseCluster(134): Setting Master Port to random. 2023-08-10T06:32:07,857 DEBUG [Listener at localhost/36659 {}] hbase.LocalHBaseCluster(141): Setting RegionServer Port to random. 2023-08-10T06:32:07,857 DEBUG [Listener at localhost/36659 {}] hbase.LocalHBaseCluster(151): Setting RS InfoServer Port to random. 2023-08-10T06:32:07,857 DEBUG [Listener at localhost/36659 {}] hbase.LocalHBaseCluster(159): Setting Master InfoServer Port to random. 2023-08-10T06:32:07,858 INFO [Listener at localhost/36659 {}] client.ConnectionUtils(128): master/jenkins-hbase3:0 server-side Connection retries=45 2023-08-10T06:32:07,858 INFO [Listener at localhost/36659 {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:07,858 INFO [Listener at localhost/36659 {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:07,858 INFO [Listener at localhost/36659 {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2023-08-10T06:32:07,858 INFO [Listener at localhost/36659 {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:07,858 INFO [Listener at localhost/36659 {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2023-08-10T06:32:07,858 INFO [Listener at localhost/36659 {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService 2023-08-10T06:32:07,858 INFO [Listener at localhost/36659 {}] ipc.NettyRpcServer(177): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2023-08-10T06:32:07,859 INFO [Listener at localhost/36659 {}] ipc.NettyRpcServer(144): Bind to /172.31.12.81:38291 2023-08-10T06:32:07,860 INFO [Listener at localhost/36659 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:07,861 INFO [Listener at localhost/36659 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:07,862 INFO [Listener at localhost/36659 {}] zookeeper.RecoverableZooKeeper(142): Process identifier=master:38291 connecting to ZooKeeper ensemble=127.0.0.1:52413 2023-08-10T06:32:07,865 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:382910x0, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2023-08-10T06:32:07,865 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(630): master:38291-0x10060ee705b0004 connected 2023-08-10T06:32:07,877 DEBUG [Listener at localhost/36659 {}] zookeeper.ZKUtil(113): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/master 2023-08-10T06:32:07,878 DEBUG [Listener at localhost/36659 {}] zookeeper.ZKUtil(113): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/running 2023-08-10T06:32:07,879 DEBUG [Listener at localhost/36659 {}] zookeeper.ZKUtil(113): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/acl 2023-08-10T06:32:07,879 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38291 2023-08-10T06:32:07,879 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38291 2023-08-10T06:32:07,879 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38291 2023-08-10T06:32:07,880 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38291 2023-08-10T06:32:07,880 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38291 2023-08-10T06:32:07,883 INFO [Listener at localhost/36659 {}] http.HttpServer(973): Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter) 2023-08-10T06:32:07,883 INFO [Listener at localhost/36659 {}] http.HttpServer(973): Added global filter 'clickjackingprevention' (class=org.apache.hadoop.hbase.http.ClickjackingPreventionFilter) 2023-08-10T06:32:07,883 INFO [Listener at localhost/36659 {}] http.HttpServer(973): Added global filter 'securityheaders' (class=org.apache.hadoop.hbase.http.SecurityHeadersFilter) 2023-08-10T06:32:07,883 INFO [Listener at localhost/36659 {}] http.HttpServer(952): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context master 2023-08-10T06:32:07,884 INFO [Listener at localhost/36659 {}] http.HttpServer(959): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2023-08-10T06:32:07,884 INFO [Listener at localhost/36659 {}] http.HttpServer(959): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2023-08-10T06:32:07,884 INFO [Listener at localhost/36659 {}] http.HttpServer(797): ASYNC_PROFILER_HOME environment variable and async.profiler.home system property not specified. Disabling /prof endpoint. 2023-08-10T06:32:07,886 INFO [Listener at localhost/36659 {}] http.HttpServer(1219): Jetty bound to port 33917 2023-08-10T06:32:07,886 INFO [Listener at localhost/36659 {}] server.Server(375): jetty-9.4.50.v20221201; built: 2022-12-01T22:07:03.915Z; git: da9a0b30691a45daf90a9f17b5defa2f1434f882; jvm 1.8.0_362-b09 2023-08-10T06:32:07,887 INFO [Listener at localhost/36659 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:07,887 INFO [Listener at localhost/36659 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.s.ServletContextHandler@45ff7847{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop.log.dir/,AVAILABLE} 2023-08-10T06:32:07,888 INFO [Listener at localhost/36659 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:07,888 INFO [Listener at localhost/36659 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.s.ServletContextHandler@3f9b45fb{static,/static,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/static/,AVAILABLE} 2023-08-10T06:32:07,901 INFO [Listener at localhost/36659 {}] webapp.StandardDescriptorProcessor(277): NO JSP Support for /, did not find org.apache.hbase.thirdparty.org.eclipse.jetty.jsp.JettyJspServlet 2023-08-10T06:32:07,903 INFO [Listener at localhost/36659 {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2023-08-10T06:32:07,903 INFO [Listener at localhost/36659 {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2023-08-10T06:32:07,903 INFO [Listener at localhost/36659 {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2023-08-10T06:32:07,904 INFO [Listener at localhost/36659 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:07,983 INFO [Listener at localhost/36659 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.w.WebAppContext@20c7d6e9{master,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/master/,AVAILABLE}{file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/master} 2023-08-10T06:32:07,984 INFO [Listener at localhost/36659 {}] server.AbstractConnector(333): Started ServerConnector@5443d879{HTTP/1.1, (http/1.1)}{0.0.0.0:33917} 2023-08-10T06:32:07,984 INFO [Listener at localhost/36659 {}] server.Server(415): Started @14241ms 2023-08-10T06:32:07,984 INFO [Listener at localhost/36659 {}] master.HMaster(484): hbase.rootdir=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa, hbase.cluster.distributed=false 2023-08-10T06:32:07,998 INFO [Listener at localhost/36659 {}] client.ConnectionUtils(128): regionserver/jenkins-hbase3:0 server-side Connection retries=45 2023-08-10T06:32:07,998 INFO [Listener at localhost/36659 {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:07,998 INFO [Listener at localhost/36659 {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:07,998 INFO [Listener at localhost/36659 {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2023-08-10T06:32:07,998 INFO [Listener at localhost/36659 {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2023-08-10T06:32:07,998 INFO [Listener at localhost/36659 {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2023-08-10T06:32:07,998 INFO [Listener at localhost/36659 {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService 2023-08-10T06:32:07,999 INFO [Listener at localhost/36659 {}] ipc.NettyRpcServer(177): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2023-08-10T06:32:08,000 INFO [Listener at localhost/36659 {}] ipc.NettyRpcServer(144): Bind to /172.31.12.81:39491 2023-08-10T06:32:08,000 INFO [Listener at localhost/36659 {}] hfile.BlockCacheFactory(127): Allocating BlockCache size=782.40 MB, blockSize=64 KB 2023-08-10T06:32:08,001 DEBUG [Listener at localhost/36659 {}] mob.MobFileCache(120): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2023-08-10T06:32:08,002 INFO [Listener at localhost/36659 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:08,004 INFO [Listener at localhost/36659 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:08,005 INFO [Listener at localhost/36659 {}] zookeeper.RecoverableZooKeeper(142): Process identifier=regionserver:39491 connecting to ZooKeeper ensemble=127.0.0.1:52413 2023-08-10T06:32:08,008 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:394910x0, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2023-08-10T06:32:08,009 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(630): regionserver:39491-0x10060ee705b0005 connected 2023-08-10T06:32:08,009 DEBUG [Listener at localhost/36659 {}] zookeeper.ZKUtil(113): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/master 2023-08-10T06:32:08,010 DEBUG [Listener at localhost/36659 {}] zookeeper.ZKUtil(113): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/running 2023-08-10T06:32:08,011 DEBUG [Listener at localhost/36659 {}] zookeeper.ZKUtil(113): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/acl 2023-08-10T06:32:08,014 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39491 2023-08-10T06:32:08,014 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39491 2023-08-10T06:32:08,014 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39491 2023-08-10T06:32:08,018 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39491 2023-08-10T06:32:08,018 DEBUG [Listener at localhost/36659 {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39491 2023-08-10T06:32:08,021 INFO [Listener at localhost/36659 {}] http.HttpServer(973): Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter) 2023-08-10T06:32:08,021 INFO [Listener at localhost/36659 {}] http.HttpServer(973): Added global filter 'clickjackingprevention' (class=org.apache.hadoop.hbase.http.ClickjackingPreventionFilter) 2023-08-10T06:32:08,021 INFO [Listener at localhost/36659 {}] http.HttpServer(973): Added global filter 'securityheaders' (class=org.apache.hadoop.hbase.http.SecurityHeadersFilter) 2023-08-10T06:32:08,022 INFO [Listener at localhost/36659 {}] http.HttpServer(952): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context regionserver 2023-08-10T06:32:08,022 INFO [Listener at localhost/36659 {}] http.HttpServer(959): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2023-08-10T06:32:08,022 INFO [Listener at localhost/36659 {}] http.HttpServer(959): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2023-08-10T06:32:08,022 INFO [Listener at localhost/36659 {}] http.HttpServer(797): ASYNC_PROFILER_HOME environment variable and async.profiler.home system property not specified. Disabling /prof endpoint. 2023-08-10T06:32:08,022 INFO [Listener at localhost/36659 {}] http.HttpServer(1219): Jetty bound to port 38915 2023-08-10T06:32:08,023 INFO [Listener at localhost/36659 {}] server.Server(375): jetty-9.4.50.v20221201; built: 2022-12-01T22:07:03.915Z; git: da9a0b30691a45daf90a9f17b5defa2f1434f882; jvm 1.8.0_362-b09 2023-08-10T06:32:08,033 INFO [Listener at localhost/36659 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:08,033 INFO [Listener at localhost/36659 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.s.ServletContextHandler@eb06edd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop.log.dir/,AVAILABLE} 2023-08-10T06:32:08,033 INFO [Listener at localhost/36659 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:08,033 INFO [Listener at localhost/36659 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.s.ServletContextHandler@487b472a{static,/static,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/static/,AVAILABLE} 2023-08-10T06:32:08,042 INFO [Listener at localhost/36659 {}] webapp.StandardDescriptorProcessor(277): NO JSP Support for /, did not find org.apache.hbase.thirdparty.org.eclipse.jetty.jsp.JettyJspServlet 2023-08-10T06:32:08,043 INFO [Listener at localhost/36659 {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2023-08-10T06:32:08,044 INFO [Listener at localhost/36659 {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2023-08-10T06:32:08,044 INFO [Listener at localhost/36659 {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2023-08-10T06:32:08,045 INFO [Listener at localhost/36659 {}] http.SecurityHeadersFilter(48): Added security headers filter 2023-08-10T06:32:08,046 INFO [Listener at localhost/36659 {}] handler.ContextHandler(921): Started o.a.h.t.o.e.j.w.WebAppContext@2695d275{regionserver,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/regionserver/,AVAILABLE}{file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/regionserver} 2023-08-10T06:32:08,047 INFO [Listener at localhost/36659 {}] server.AbstractConnector(333): Started ServerConnector@67bd5ab4{HTTP/1.1, (http/1.1)}{0.0.0.0:38915} 2023-08-10T06:32:08,047 INFO [Listener at localhost/36659 {}] server.Server(415): Started @14304ms 2023-08-10T06:32:08,049 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] server.Server(375): jetty-9.4.50.v20221201; built: 2022-12-01T22:07:03.915Z; git: da9a0b30691a45daf90a9f17b5defa2f1434f882; jvm 1.8.0_362-b09 2023-08-10T06:32:08,053 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] server.AbstractConnector(333): Started ServerConnector@2e1f0112{HTTP/1.1, (http/1.1)}{0.0.0.0:37999} 2023-08-10T06:32:08,053 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] server.Server(415): Started @14310ms 2023-08-10T06:32:08,053 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(2387): Adding backup master ZNode /2/backup-masters/jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:08,056 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2/backup-masters 2023-08-10T06:32:08,056 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2/backup-masters 2023-08-10T06:32:08,057 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on existing znode=/2/backup-masters/jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:08,060 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/2/master 2023-08-10T06:32:08,060 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,060 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/2/master 2023-08-10T06:32:08,061 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,062 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on existing znode=/2/master 2023-08-10T06:32:08,063 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /2/backup-masters/jenkins-hbase3.apache.org,38291,1691649127857 from backup master directory 2023-08-10T06:32:08,064 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;jenkins-hbase3:38291 2023-08-10T06:32:08,064 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on existing znode=/2/master 2023-08-10T06:32:08,065 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/2/backup-masters/jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:08,065 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2/backup-masters 2023-08-10T06:32:08,065 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2/backup-masters 2023-08-10T06:32:08,065 WARN [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2023-08-10T06:32:08,065 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:08,086 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] util.FSUtils(610): Created cluster ID file at hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/hbase.id with ID: ddd39841-4c42-4031-82a9-18a84f6a72c6 2023-08-10T06:32:08,104 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:08,107 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,107 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,122 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegion(355): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:08,123 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2023-08-10T06:32:08,124 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2023-08-10T06:32:08,140 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(7093): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store 2023-08-10T06:32:08,155 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(888): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:08,155 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1668): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2023-08-10T06:32:08,155 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1690): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:32:08,155 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1711): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:32:08,155 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1778): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2023-08-10T06:32:08,155 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1788): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:32:08,155 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1909): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:32:08,155 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1622): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2023-08-10T06:32:08,160 WARN [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegion(232): failed to clean up initializing flag: hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/.initializing 2023-08-10T06:32:08,160 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegion(202): WALDir=hdfs://localhost:44455/tmp/wal/MasterData/WALs/jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:08,161 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2023-08-10T06:32:08,164 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.AbstractFSWAL(491): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=jenkins-hbase3.apache.org%2C38291%2C1691649127857, suffix=, logDir=hdfs://localhost:44455/tmp/wal/MasterData/WALs/jenkins-hbase3.apache.org,38291,1691649127857, archiveDir=hdfs://localhost:44455/tmp/wal/MasterData/oldWALs, maxLogs=10 2023-08-10T06:32:08,177 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /tmp/wal/MasterData/WALs/jenkins-hbase3.apache.org,38291,1691649127857/jenkins-hbase3.apache.org%2C38291%2C1691649127857.1691649128164, exclude list is [], retry=0 2023-08-10T06:32:08,181 DEBUG [RS-EventLoopGroup-6-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42505,DS-9386f1a7-3604-4ae3-9aed-a633fa75b158,DISK] 2023-08-10T06:32:08,185 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.AbstractFSWAL(832): New WAL /tmp/wal/MasterData/WALs/jenkins-hbase3.apache.org,38291,1691649127857/jenkins-hbase3.apache.org%2C38291%2C1691649127857.1691649128164 2023-08-10T06:32:08,185 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:42505,DS-9386f1a7-3604-4ae3-9aed-a633fa75b158,DISK]] 2023-08-10T06:32:08,185 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(7254): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:08,186 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(888): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:08,186 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(7296): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,186 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(7299): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,188 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,190 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2023-08-10T06:32:08,190 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:08,191 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(324): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:08,191 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,193 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2023-08-10T06:32:08,193 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:08,194 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(324): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:08,194 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,196 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2023-08-10T06:32:08,196 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:08,197 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(324): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:08,197 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,198 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2023-08-10T06:32:08,199 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:08,199 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(324): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:08,201 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,201 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,202 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/tmp/wal/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,206 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2023-08-10T06:32:08,208 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1076): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2023-08-10T06:32:08,218 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:44455/tmp/wal/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:08,219 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(1093): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=9615464160, jitterRate=-0.10449011623859406}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2023-08-10T06:32:08,219 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] regionserver.HRegion(992): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2023-08-10T06:32:08,219 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2023-08-10T06:32:08,224 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@62bb7010, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:08,225 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(859): No meta location available on zookeeper, skip migrating... 2023-08-10T06:32:08,225 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2023-08-10T06:32:08,226 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(564): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2023-08-10T06:32:08,226 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2023-08-10T06:32:08,226 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(584): Recovered RegionProcedureStore lease in 0 msec 2023-08-10T06:32:08,227 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(598): Loaded RegionProcedureStore in 0 msec 2023-08-10T06:32:08,227 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2023-08-10T06:32:08,229 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2023-08-10T06:32:08,230 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Unable to get data of znode /2/balancer because node does not exist (not necessarily an error) 2023-08-10T06:32:08,232 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /2/balancer already deleted, retry=false 2023-08-10T06:32:08,232 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(150): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2023-08-10T06:32:08,232 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Unable to get data of znode /2/normalizer because node does not exist (not necessarily an error) 2023-08-10T06:32:08,234 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /2/normalizer already deleted, retry=false 2023-08-10T06:32:08,234 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2023-08-10T06:32:08,235 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Unable to get data of znode /2/switch/split because node does not exist (not necessarily an error) 2023-08-10T06:32:08,236 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /2/switch/split already deleted, retry=false 2023-08-10T06:32:08,237 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Unable to get data of znode /2/switch/merge because node does not exist (not necessarily an error) 2023-08-10T06:32:08,238 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /2/switch/merge already deleted, retry=false 2023-08-10T06:32:08,239 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Unable to get data of znode /2/snapshot-cleanup because node does not exist (not necessarily an error) 2023-08-10T06:32:08,240 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(224): Node /2/snapshot-cleanup already deleted, retry=false 2023-08-10T06:32:08,243 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/2/running 2023-08-10T06:32:08,243 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/2/running 2023-08-10T06:32:08,243 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,243 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,243 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(803): Active/primary master=jenkins-hbase3.apache.org,38291,1691649127857, sessionid=0x10060ee705b0004, setting cluster-up flag (Was=false) 2023-08-10T06:32:08,244 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ProcedureManagerHost(66): User procedure org.apache.hadoop.hbase.backup.master.LogRollMasterProcedureManager was loaded successfully. 2023-08-10T06:32:08,246 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,246 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,252 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /2/flush-table-proc/acquired, /2/flush-table-proc/reached, /2/flush-table-proc/abort 2023-08-10T06:32:08,253 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:08,263 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,263 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,269 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /2/rolllog-proc/acquired, /2/rolllog-proc/reached, /2/rolllog-proc/abort 2023-08-10T06:32:08,270 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:08,273 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,273 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:08,278 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /2/online-snapshot/acquired, /2/online-snapshot/reached, /2/online-snapshot/abort 2023-08-10T06:32:08,279 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:08,283 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1033): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2023-08-10T06:32:08,283 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(574): slop=0.2 2023-08-10T06:32:08,284 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(285): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2023-08-10T06:32:08,284 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: jenkins-hbase3.apache.org,38291,1691649127857 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2023-08-10T06:32:08,285 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/jenkins-hbase3:0, corePoolSize=5, maxPoolSize=5 2023-08-10T06:32:08,285 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/jenkins-hbase3:0, corePoolSize=5, maxPoolSize=5 2023-08-10T06:32:08,285 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=5, maxPoolSize=5 2023-08-10T06:32:08,285 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=5, maxPoolSize=5 2023-08-10T06:32:08,285 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/jenkins-hbase3:0, corePoolSize=10, maxPoolSize=10 2023-08-10T06:32:08,285 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,285 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=2, maxPoolSize=2 2023-08-10T06:32:08,286 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,291 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1691649158291 2023-08-10T06:32:08,292 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2023-08-10T06:32:08,292 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2023-08-10T06:32:08,292 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2023-08-10T06:32:08,292 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.backup.master.BackupLogCleaner 2023-08-10T06:32:08,292 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2023-08-10T06:32:08,292 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2023-08-10T06:32:08,292 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2023-08-10T06:32:08,294 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,294 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2023-08-10T06:32:08,294 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2023-08-10T06:32:08,294 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2023-08-10T06:32:08,295 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2023-08-10T06:32:08,295 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2023-08-10T06:32:08,295 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ReadOnlyZKClient(139): Connect 0x72266fe4 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:08,296 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:08,296 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2023-08-10T06:32:08,316 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@17ab7ccb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:08,317 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.backup.BackupHFileCleaner 2023-08-10T06:32:08,322 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2023-08-10T06:32:08,322 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2023-08-10T06:32:08,322 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.large.0-1691649128322,5,FailOnTimeoutGroup] 2023-08-10T06:32:08,322 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.small.0-1691649128322,5,FailOnTimeoutGroup] 2023-08-10T06:32:08,322 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,323 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(1634): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2023-08-10T06:32:08,323 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,323 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,358 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;jenkins-hbase3:39491 2023-08-10T06:32:08,359 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(997): ClusterId : ddd39841-4c42-4031-82a9-18a84f6a72c6 2023-08-10T06:32:08,359 INFO [RS:0;jenkins-hbase3:39491 {}] procedure.ProcedureManagerHost(66): User procedure org.apache.hadoop.hbase.backup.regionserver.LogRollRegionServerProcedureManager was loaded successfully. 2023-08-10T06:32:08,359 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2023-08-10T06:32:08,362 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2023-08-10T06:32:08,362 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(43): Procedure backup-proc initializing 2023-08-10T06:32:08,367 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(45): Procedure backup-proc initialized 2023-08-10T06:32:08,367 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2023-08-10T06:32:08,371 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2023-08-10T06:32:08,372 DEBUG [RS:0;jenkins-hbase3:39491 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5df844a0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:08,378 DEBUG [RS:0;jenkins-hbase3:39491 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@194672fb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=jenkins-hbase3.apache.org/172.31.12.81:0 2023-08-10T06:32:08,378 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2023-08-10T06:32:08,378 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2023-08-10T06:32:08,378 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1079): About to register with Master. 2023-08-10T06:32:08,380 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(3023): reportForDuty to master=jenkins-hbase3.apache.org,38291,1691649127857 with isa=jenkins-hbase3.apache.org/172.31.12.81:39491, startcode=1691649127997 2023-08-10T06:32:08,380 DEBUG [RS:0;jenkins-hbase3:39491 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2023-08-10T06:32:08,387 INFO [RS-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:43895, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2023-08-10T06:32:08,389 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38291 {}] master.ServerManager(403): Registering regionserver=jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:08,393 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1705): Config from master: hbase.rootdir=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa 2023-08-10T06:32:08,393 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1705): Config from master: fs.defaultFS=hdfs://localhost:44455 2023-08-10T06:32:08,393 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1705): Config from master: hbase.master.info.port=33917 2023-08-10T06:32:08,396 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2/rs 2023-08-10T06:32:08,396 DEBUG [RS:0;jenkins-hbase3:39491 {}] zookeeper.ZKUtil(111): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on existing znode=/2/rs/jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:08,396 WARN [RS:0;jenkins-hbase3:39491 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2023-08-10T06:32:08,397 INFO [RS:0;jenkins-hbase3:39491 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2023-08-10T06:32:08,397 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(2060): logDir=hdfs://localhost:44455/tmp/wal/WALs/jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:08,397 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [jenkins-hbase3.apache.org,39491,1691649127997] 2023-08-10T06:32:08,405 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2023-08-10T06:32:08,405 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.MetricsRegionServerWrapperImpl(165): Computing regionserver metrics every 5000 milliseconds 2023-08-10T06:32:08,408 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=782.4 M, globalMemStoreLimitLowMark=743.3 M, Offheap=false 2023-08-10T06:32:08,410 INFO [RS:0;jenkins-hbase3:39491 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2023-08-10T06:32:08,410 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,414 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer$CompactionChecker(1947): CompactionChecker runs every PT1S 2023-08-10T06:32:08,416 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,416 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,416 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,416 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,416 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,417 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,417 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/jenkins-hbase3:0, corePoolSize=2, maxPoolSize=2 2023-08-10T06:32:08,417 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,417 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,417 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,417 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,417 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/jenkins-hbase3:0, corePoolSize=1, maxPoolSize=1 2023-08-10T06:32:08,418 DEBUG [RS:0;jenkins-hbase3:39491 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0, corePoolSize=3, maxPoolSize=3 2023-08-10T06:32:08,420 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,420 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,420 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,420 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,420 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,39491,1691649127997-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2023-08-10T06:32:08,434 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2023-08-10T06:32:08,434 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,39491,1691649127997-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:08,445 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.Replication(204): jenkins-hbase3.apache.org,39491,1691649127997 started 2023-08-10T06:32:08,445 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1747): Serving as jenkins-hbase3.apache.org,39491,1691649127997, RpcServer on jenkins-hbase3.apache.org/172.31.12.81:39491, sessionid=0x10060ee705b0005 2023-08-10T06:32:08,445 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2023-08-10T06:32:08,445 DEBUG [RS:0;jenkins-hbase3:39491 {}] flush.RegionServerFlushTableProcedureManager(106): Start region server flush procedure manager jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:08,445 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'jenkins-hbase3.apache.org,39491,1691649127997' 2023-08-10T06:32:08,445 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/2/flush-table-proc/abort' 2023-08-10T06:32:08,446 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/2/flush-table-proc/acquired' 2023-08-10T06:32:08,447 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2023-08-10T06:32:08,447 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(51): Procedure backup-proc starting 2023-08-10T06:32:08,447 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'jenkins-hbase3.apache.org,39491,1691649127997' 2023-08-10T06:32:08,447 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/2/rolllog-proc/abort' 2023-08-10T06:32:08,448 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/2/rolllog-proc/acquired' 2023-08-10T06:32:08,448 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.LogRollRegionServerProcedureManager(92): Started region server backup manager. 2023-08-10T06:32:08,448 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(53): Procedure backup-proc started 2023-08-10T06:32:08,448 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2023-08-10T06:32:08,448 DEBUG [RS:0;jenkins-hbase3:39491 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:08,449 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'jenkins-hbase3.apache.org,39491,1691649127997' 2023-08-10T06:32:08,449 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/2/online-snapshot/abort' 2023-08-10T06:32:08,449 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/2/online-snapshot/acquired' 2023-08-10T06:32:08,450 DEBUG [RS:0;jenkins-hbase3:39491 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2023-08-10T06:32:08,450 INFO [RS:0;jenkins-hbase3:39491 {}] quotas.RegionServerRpcQuotaManager(63): Quota support disabled 2023-08-10T06:32:08,450 INFO [RS:0;jenkins-hbase3:39491 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2023-08-10T06:32:08,551 INFO [RS:0;jenkins-hbase3:39491 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2023-08-10T06:32:08,554 INFO [RS:0;jenkins-hbase3:39491 {}] wal.AbstractFSWAL(491): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=jenkins-hbase3.apache.org%2C39491%2C1691649127997, suffix=, logDir=hdfs://localhost:44455/tmp/wal/WALs/jenkins-hbase3.apache.org,39491,1691649127997, archiveDir=hdfs://localhost:44455/tmp/wal/oldWALs, maxLogs=32 2023-08-10T06:32:08,571 DEBUG [RS:0;jenkins-hbase3:39491 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /tmp/wal/WALs/jenkins-hbase3.apache.org,39491,1691649127997/jenkins-hbase3.apache.org%2C39491%2C1691649127997.1691649128555, exclude list is [], retry=0 2023-08-10T06:32:08,578 DEBUG [RS-EventLoopGroup-6-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:34143,DS-3ea9105e-de54-44d1-a029-d02e873daf8c,DISK] 2023-08-10T06:32:08,586 INFO [RS:0;jenkins-hbase3:39491 {}] wal.AbstractFSWAL(832): New WAL /tmp/wal/WALs/jenkins-hbase3.apache.org,39491,1691649127997/jenkins-hbase3.apache.org%2C39491%2C1691649127997.1691649128555 2023-08-10T06:32:08,586 DEBUG [RS:0;jenkins-hbase3:39491 {}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:34143,DS-3ea9105e-de54-44d1-a029-d02e873daf8c,DISK]] 2023-08-10T06:32:08,728 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2023-08-10T06:32:08,729 INFO [PEWorker-1 {}] regionserver.HRegion(7075): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa 2023-08-10T06:32:09,150 DEBUG [PEWorker-1 {}] regionserver.HRegion(888): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:09,152 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2023-08-10T06:32:09,155 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2023-08-10T06:32:09,156 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:09,156 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:09,156 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2023-08-10T06:32:09,158 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2023-08-10T06:32:09,158 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:09,159 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:09,159 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2023-08-10T06:32:09,161 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2023-08-10T06:32:09,161 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:09,161 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:09,162 DEBUG [PEWorker-1 {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740 2023-08-10T06:32:09,163 DEBUG [PEWorker-1 {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740 2023-08-10T06:32:09,164 DEBUG [PEWorker-1 {}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/tmp/wal/data/hbase/meta/1588230740 2023-08-10T06:32:09,171 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2023-08-10T06:32:09,172 DEBUG [PEWorker-1 {}] regionserver.HRegion(1076): writing seq id for 1588230740 2023-08-10T06:32:09,176 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:44455/tmp/wal/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:09,176 INFO [PEWorker-1 {}] regionserver.HRegion(1093): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=10040887520, jitterRate=-0.0648694783449173}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2023-08-10T06:32:09,176 DEBUG [PEWorker-1 {}] regionserver.HRegion(992): Region open journal for 1588230740: 2023-08-10T06:32:09,176 DEBUG [PEWorker-1 {}] regionserver.HRegion(1668): Closing 1588230740, disabling compactions & flushes 2023-08-10T06:32:09,176 INFO [PEWorker-1 {}] regionserver.HRegion(1690): Closing region hbase:meta,,1.1588230740 2023-08-10T06:32:09,176 DEBUG [PEWorker-1 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2023-08-10T06:32:09,177 DEBUG [PEWorker-1 {}] regionserver.HRegion(1778): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2023-08-10T06:32:09,177 DEBUG [PEWorker-1 {}] regionserver.HRegion(1788): Updates disabled for region hbase:meta,,1.1588230740 2023-08-10T06:32:09,177 INFO [PEWorker-1 {}] regionserver.HRegion(1909): Closed hbase:meta,,1.1588230740 2023-08-10T06:32:09,177 DEBUG [PEWorker-1 {}] regionserver.HRegion(1622): Region close journal for 1588230740: 2023-08-10T06:32:09,178 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2023-08-10T06:32:09,178 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2023-08-10T06:32:09,179 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2023-08-10T06:32:09,181 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2023-08-10T06:32:09,183 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(262): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2023-08-10T06:32:09,308 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2023-08-10T06:32:09,309 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:09,315 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2023-08-10T06:32:09,315 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:09,316 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:32:09,316 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.backup.BackupObserver Metrics about HBase RegionObservers 2023-08-10T06:32:09,317 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2023-08-10T06:32:09,317 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2023-08-10T06:32:09,317 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2023-08-10T06:32:09,317 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:09,333 DEBUG [jenkins-hbase3:38291 {}] assignment.AssignmentManager(2303): Processing assignQueue; systemServersCount=1, allServersCount=1 2023-08-10T06:32:09,334 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:09,337 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as jenkins-hbase3.apache.org,39491,1691649127997, state=OPENING 2023-08-10T06:32:09,339 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2023-08-10T06:32:09,342 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:09,342 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:09,342 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=jenkins-hbase3.apache.org,39491,1691649127997}] 2023-08-10T06:32:09,343 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /2/meta-region-server: CHANGED 2023-08-10T06:32:09,343 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /2/meta-region-server: CHANGED 2023-08-10T06:32:09,498 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(721): New admin connection to jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:09,499 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=AdminService, sasl=false 2023-08-10T06:32:09,502 INFO [RS-EventLoopGroup-6-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:56954, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2023-08-10T06:32:09,509 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(130): Open hbase:meta,,1.1588230740 2023-08-10T06:32:09,509 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2023-08-10T06:32:09,509 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2023-08-10T06:32:09,511 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(491): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=jenkins-hbase3.apache.org%2C39491%2C1691649127997.meta, suffix=.meta, logDir=hdfs://localhost:44455/tmp/wal/WALs/jenkins-hbase3.apache.org,39491,1691649127997, archiveDir=hdfs://localhost:44455/tmp/wal/oldWALs, maxLogs=32 2023-08-10T06:32:09,524 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /tmp/wal/WALs/jenkins-hbase3.apache.org,39491,1691649127997/jenkins-hbase3.apache.org%2C39491%2C1691649127997.meta.1691649129512.meta, exclude list is [], retry=0 2023-08-10T06:32:09,529 DEBUG [RS-EventLoopGroup-6-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37669,DS-660e422a-4614-4be2-bdae-34ba7271ed00,DISK] 2023-08-10T06:32:09,531 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(832): New WAL /tmp/wal/WALs/jenkins-hbase3.apache.org,39491,1691649127997/jenkins-hbase3.apache.org%2C39491%2C1691649127997.meta.1691649129512.meta 2023-08-10T06:32:09,532 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:37669,DS-660e422a-4614-4be2-bdae-34ba7271ed00,DISK]] 2023-08-10T06:32:09,532 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7254): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:09,532 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:09,532 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2023-08-10T06:32:09,532 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7955): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2023-08-10T06:32:09,532 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(393): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2023-08-10T06:32:09,533 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2023-08-10T06:32:09,533 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(888): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:09,533 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7296): checking encryption for 1588230740 2023-08-10T06:32:09,533 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7299): checking classloading for 1588230740 2023-08-10T06:32:09,534 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2023-08-10T06:32:09,536 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2023-08-10T06:32:09,536 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:09,536 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:09,536 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2023-08-10T06:32:09,537 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2023-08-10T06:32:09,538 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:09,538 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:09,538 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2023-08-10T06:32:09,540 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2023-08-10T06:32:09,540 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:09,540 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(324): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2023-08-10T06:32:09,541 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740 2023-08-10T06:32:09,542 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740 2023-08-10T06:32:09,544 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/tmp/wal/data/hbase/meta/1588230740 2023-08-10T06:32:09,546 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2023-08-10T06:32:09,548 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1076): writing seq id for 1588230740 2023-08-10T06:32:09,549 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=10006311200, jitterRate=-0.06808964908123016}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2023-08-10T06:32:09,549 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(992): Region open journal for 1588230740: 2023-08-10T06:32:09,551 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2552): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1691649129498 2023-08-10T06:32:09,556 DEBUG [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2579): Finished post open deploy task for hbase:meta,,1.1588230740 2023-08-10T06:32:09,557 INFO [RS_OPEN_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(158): Opened hbase:meta,,1.1588230740 2023-08-10T06:32:09,557 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:09,559 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as jenkins-hbase3.apache.org,39491,1691649127997, state=OPEN 2023-08-10T06:32:09,561 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/2/meta-region-server 2023-08-10T06:32:09,561 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/2/meta-region-server 2023-08-10T06:32:09,561 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /2/meta-region-server: CHANGED 2023-08-10T06:32:09,561 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /2/meta-region-server: CHANGED 2023-08-10T06:32:09,564 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=3, resume processing ppid=2 2023-08-10T06:32:09,564 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=jenkins-hbase3.apache.org,39491,1691649127997 in 219 msec 2023-08-10T06:32:09,567 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=2, resume processing ppid=1 2023-08-10T06:32:09,567 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 385 msec 2023-08-10T06:32:09,570 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 1.2870 sec 2023-08-10T06:32:09,570 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(1062): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1691649129570, completionTime=-1 2023-08-10T06:32:09,571 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.ServerManager(830): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2023-08-10T06:32:09,571 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] assignment.AssignmentManager(1629): Joining cluster... 2023-08-10T06:32:09,574 DEBUG [hconnection-0x4624cea2-shared-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:09,578 INFO [RS-EventLoopGroup-6-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:56966, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:09,580 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] assignment.AssignmentManager(1641): Number of RegionServers=1 2023-08-10T06:32:09,580 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1691649189580 2023-08-10T06:32:09,580 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1691649249580 2023-08-10T06:32:09,580 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] assignment.AssignmentManager(1648): Joined the cluster in 9 msec 2023-08-10T06:32:09,589 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,38291,1691649127857-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:09,589 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,38291,1691649127857-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:09,589 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,38291,1691649127857-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:09,590 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-jenkins-hbase3:38291, period=300000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:09,590 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2023-08-10T06:32:09,590 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2023-08-10T06:32:09,590 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(2367): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2023-08-10T06:32:09,592 DEBUG [master/jenkins-hbase3:0.Chore.1 {}] janitor.CatalogJanitor(175): 2023-08-10T06:32:09,592 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1033): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2023-08-10T06:32:09,594 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:09,594 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:09,595 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:09,936 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2023-08-10T06:32:09,937 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2023-08-10T06:32:10,015 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => 86e6ed14dc5c14940dff214bf60e7e35, NAME => 'hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa 2023-08-10T06:32:10,031 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(888): Instantiated hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:10,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1668): Closing 86e6ed14dc5c14940dff214bf60e7e35, disabling compactions & flushes 2023-08-10T06:32:10,032 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1690): Closing region hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:32:10,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:32:10,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. after waiting 0 ms 2023-08-10T06:32:10,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:32:10,032 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1909): Closed hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:32:10,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1622): Region close journal for 86e6ed14dc5c14940dff214bf60e7e35: 2023-08-10T06:32:10,033 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:10,034 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1691649130033"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649130033"}]},"ts":"1691649130033"} 2023-08-10T06:32:10,037 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:10,038 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:10,039 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649130038"}]},"ts":"1691649130038"} 2023-08-10T06:32:10,042 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2023-08-10T06:32:10,047 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=86e6ed14dc5c14940dff214bf60e7e35, ASSIGN}] 2023-08-10T06:32:10,049 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=86e6ed14dc5c14940dff214bf60e7e35, ASSIGN 2023-08-10T06:32:10,050 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(262): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=86e6ed14dc5c14940dff214bf60e7e35, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,39491,1691649127997; forceNewPlan=false, retain=false 2023-08-10T06:32:10,201 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=86e6ed14dc5c14940dff214bf60e7e35, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:10,204 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 86e6ed14dc5c14940dff214bf60e7e35, server=jenkins-hbase3.apache.org,39491,1691649127997}] 2023-08-10T06:32:10,363 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(130): Open hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:32:10,364 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7254): Opening region: {ENCODED => 86e6ed14dc5c14940dff214bf60e7e35, NAME => 'hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:10,364 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:10,364 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:32:10,365 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(888): Instantiated hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:10,365 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7296): checking encryption for 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:32:10,365 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7299): checking classloading for 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:32:10,367 INFO [StoreOpener-86e6ed14dc5c14940dff214bf60e7e35-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:32:10,369 INFO [StoreOpener-86e6ed14dc5c14940dff214bf60e7e35-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 86e6ed14dc5c14940dff214bf60e7e35 columnFamilyName info 2023-08-10T06:32:10,369 DEBUG [StoreOpener-86e6ed14dc5c14940dff214bf60e7e35-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:10,370 INFO [StoreOpener-86e6ed14dc5c14940dff214bf60e7e35-1 {}] regionserver.HStore(324): Store=86e6ed14dc5c14940dff214bf60e7e35/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:10,372 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:32:10,372 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:32:10,373 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:44455/tmp/wal/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:32:10,378 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1076): writing seq id for 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:32:10,381 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:44455/tmp/wal/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:10,382 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1093): Opened 86e6ed14dc5c14940dff214bf60e7e35; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11227501760, jitterRate=0.045642584562301636}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2023-08-10T06:32:10,382 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(992): Region open journal for 86e6ed14dc5c14940dff214bf60e7e35: 2023-08-10T06:32:10,384 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2552): Post open deploy tasks for hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35., pid=6, masterSystemTime=1691649130357 2023-08-10T06:32:10,387 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2579): Finished post open deploy task for hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:32:10,387 INFO [RS_OPEN_PRIORITY_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(158): Opened hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:32:10,388 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=86e6ed14dc5c14940dff214bf60e7e35, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:32:10,393 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=6, resume processing ppid=5 2023-08-10T06:32:10,393 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 86e6ed14dc5c14940dff214bf60e7e35, server=jenkins-hbase3.apache.org,39491,1691649127997 in 186 msec 2023-08-10T06:32:10,397 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=5, resume processing ppid=4 2023-08-10T06:32:10,397 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1414): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=86e6ed14dc5c14940dff214bf60e7e35, ASSIGN in 346 msec 2023-08-10T06:32:10,398 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:10,398 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649130398"}]},"ts":"1691649130398"} 2023-08-10T06:32:10,401 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2023-08-10T06:32:10,404 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:10,407 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 814 msec 2023-08-10T06:32:10,494 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/namespace 2023-08-10T06:32:10,495 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:10,495 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/2/namespace 2023-08-10T06:32:10,495 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:32:10,502 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1033): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2023-08-10T06:32:10,516 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2/namespace 2023-08-10T06:32:10,522 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 19 msec 2023-08-10T06:32:10,525 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1033): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2023-08-10T06:32:10,540 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2/namespace 2023-08-10T06:32:10,545 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 19 msec 2023-08-10T06:32:10,551 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/2/namespace/default 2023-08-10T06:32:10,553 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/2/namespace/hbase 2023-08-10T06:32:10,554 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(1192): Master has completed initialization 2.488sec 2023-08-10T06:32:10,554 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2023-08-10T06:32:10,554 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2023-08-10T06:32:10,554 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2023-08-10T06:32:10,554 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2023-08-10T06:32:10,554 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2023-08-10T06:32:10,554 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,38291,1691649127857-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2023-08-10T06:32:10,554 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=jenkins-hbase3.apache.org,38291,1691649127857-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2023-08-10T06:32:10,559 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster {}] master.HMaster(1295): Balancer post startup initialization complete, took 0 seconds 2023-08-10T06:32:10,559 INFO [master/jenkins-hbase3:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2023-08-10T06:32:10,648 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2023-08-10T06:32:10,652 DEBUG [Listener at localhost/36659 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x1cbcc976 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:10,665 DEBUG [Listener at localhost/36659 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2850837, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:10,677 DEBUG [hconnection-0x377f85cf-shared-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:10,691 INFO [RS-EventLoopGroup-6-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:45312, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:10,695 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:32:10,695 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(2786): Starting mini mapreduce cluster... 2023-08-10T06:32:10,695 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/test.cache.data in system properties and HBase conf 2023-08-10T06:32:10,696 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.tmp.dir in system properties and HBase conf 2023-08-10T06:32:10,696 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.log.dir in system properties and HBase conf 2023-08-10T06:32:10,696 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.local.dir in system properties and HBase conf 2023-08-10T06:32:10,696 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.temp.dir in system properties and HBase conf 2023-08-10T06:32:10,696 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2023-08-10T06:32:10,696 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:32:10,697 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:32:10,697 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.nodemanager.log-dirs in system properties and HBase conf 2023-08-10T06:32:10,697 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:32:10,697 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2023-08-10T06:32:10,697 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2023-08-10T06:32:10,698 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:32:10,698 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:32:10,698 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2023-08-10T06:32:10,699 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/nfs.dump.dir in system properties and HBase conf 2023-08-10T06:32:10,699 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/java.io.tmpdir in system properties and HBase conf 2023-08-10T06:32:10,699 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:32:10,699 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2023-08-10T06:32:10,699 INFO [Listener at localhost/36659 {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2023-08-10T06:32:11,972 WARN [Thread-467 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:12,330 INFO [Thread-467 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:12,337 INFO [Thread-467 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/2.10.2/hadoop-yarn-common-2.10.2.jar!/webapps/jobhistory to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/java.io.tmpdir/Jetty_jenkins.hbase3_apache_org_33903_jobhistory____.rryoey/webapp 2023-08-10T06:32:12,541 INFO [Thread-467 {}] container.GuiceComponentProviderFactory(107): Registering org.apache.hadoop.mapreduce.v2.hs.webapp.HsWebServices as a root resource class 2023-08-10T06:32:12,542 INFO [Thread-467 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.mapreduce.v2.hs.webapp.JAXBContextResolver as a provider class 2023-08-10T06:32:12,542 INFO [Thread-467 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2023-08-10T06:32:12,546 INFO [Thread-467 {}] application.WebApplicationImpl(791): Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM' 2023-08-10T06:32:12,638 INFO [Thread-467 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.mapreduce.v2.hs.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:12,746 INFO [Thread-467 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:13,034 INFO [Thread-467 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.mapreduce.v2.hs.webapp.HsWebServices to GuiceManagedComponentProvider with the scope "PerRequest" 2023-08-10T06:32:13,067 INFO [Thread-467 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:33903 2023-08-10T06:32:13,465 WARN [Listener at jenkins-hbase3.apache.org/34767 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:13,550 WARN [Listener at jenkins-hbase3.apache.org/34767 {}] servlet.GuiceFilter(87): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2023-08-10T06:32:13,551 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:13,557 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/2.10.2/hadoop-yarn-common-2.10.2.jar!/webapps/cluster to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/java.io.tmpdir/Jetty_jenkins.hbase3_apache_org_42743_cluster____4qs417/webapp 2023-08-10T06:32:13,629 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver as a provider class 2023-08-10T06:32:13,629 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] container.GuiceComponentProviderFactory(107): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices as a root resource class 2023-08-10T06:32:13,629 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2023-08-10T06:32:13,630 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] application.WebApplicationImpl(791): Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM' 2023-08-10T06:32:13,638 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:13,839 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:14,002 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:14,015 INFO [Listener at jenkins-hbase3.apache.org/34767 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:42743 2023-08-10T06:32:14,434 WARN [Listener at jenkins-hbase3.apache.org/33649 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:14,458 WARN [Listener at jenkins-hbase3.apache.org/33649 {}] servlet.GuiceFilter(87): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2023-08-10T06:32:14,458 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:14,464 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/2.10.2/hadoop-yarn-common-2.10.2.jar!/webapps/node to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/java.io.tmpdir/Jetty_jenkins.hbase3_apache_org_39599_node____.lbd3oy/webapp 2023-08-10T06:32:14,514 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] container.GuiceComponentProviderFactory(107): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2023-08-10T06:32:14,515 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2023-08-10T06:32:14,515 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2023-08-10T06:32:14,515 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] application.WebApplicationImpl(791): Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM' 2023-08-10T06:32:14,522 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:14,544 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:14,615 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:14,626 INFO [Listener at jenkins-hbase3.apache.org/33649 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:39599 2023-08-10T06:32:14,795 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:14,803 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] servlet.GuiceFilter(87): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2023-08-10T06:32:14,804 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:14,815 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/2.10.2/hadoop-yarn-common-2.10.2.jar!/webapps/node to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/java.io.tmpdir/Jetty_jenkins.hbase3_apache_org_44877_node____bbg4af/webapp 2023-08-10T06:32:14,890 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] container.GuiceComponentProviderFactory(107): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2023-08-10T06:32:14,891 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2023-08-10T06:32:14,891 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2023-08-10T06:32:14,891 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] application.WebApplicationImpl(791): Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM' 2023-08-10T06:32:14,903 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:14,929 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:15,032 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:15,040 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:44877 2023-08-10T06:32:15,085 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(2821): Mini mapreduce cluster started 2023-08-10T06:32:15,086 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] backup.TestBackupBase(301): ROOTDIR hdfs://localhost:33549/backupUT 2023-08-10T06:32:15,102 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] backup.TestBackupBase(306): REMOTE ROOTDIR hdfs://localhost:44455/backupUT 2023-08-10T06:32:15,119 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:15,123 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:56882, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:32:15,133 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$17(3366): Client=jenkins//172.31.12.81 creating {NAME => 'ns1'} 2023-08-10T06:32:15,136 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=9, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=ns1 2023-08-10T06:32:15,152 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=9 2023-08-10T06:32:15,165 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/namespace 2023-08-10T06:32:15,171 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=9, state=SUCCESS; CreateNamespaceProcedure, namespace=ns1 in 35 msec 2023-08-10T06:32:15,258 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=9 2023-08-10T06:32:15,260 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$17(3366): Client=jenkins//172.31.12.81 creating {NAME => 'ns2'} 2023-08-10T06:32:15,261 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=10, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=ns2 2023-08-10T06:32:15,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=10 2023-08-10T06:32:15,273 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/namespace 2023-08-10T06:32:15,278 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1414): Finished pid=10, state=SUCCESS; CreateNamespaceProcedure, namespace=ns2 in 16 msec 2023-08-10T06:32:15,386 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=10 2023-08-10T06:32:15,388 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$17(3366): Client=jenkins//172.31.12.81 creating {NAME => 'ns3'} 2023-08-10T06:32:15,389 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=11, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=ns3 2023-08-10T06:32:15,394 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=11 2023-08-10T06:32:15,400 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/namespace 2023-08-10T06:32:15,405 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=11, state=SUCCESS; CreateNamespaceProcedure, namespace=ns3 in 15 msec 2023-08-10T06:32:15,496 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=11 2023-08-10T06:32:15,497 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$17(3366): Client=jenkins//172.31.12.81 creating {NAME => 'ns4'} 2023-08-10T06:32:15,498 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=12, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=ns4 2023-08-10T06:32:15,503 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=12 2023-08-10T06:32:15,509 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/namespace 2023-08-10T06:32:15,518 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=12, state=SUCCESS; CreateNamespaceProcedure, namespace=ns4 in 20 msec 2023-08-10T06:32:15,606 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=12 2023-08-10T06:32:15,623 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$4(2331): Client=jenkins//172.31.12.81 create 'test-1691649135102', {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:15,626 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=13, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=test-1691649135102 2023-08-10T06:32:15,629 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=13, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=test-1691649135102 execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:15,629 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(708): Client=jenkins//172.31.12.81 procedure request for creating table: namespace: "default" qualifier: "test-1691649135102" procId is: 13 2023-08-10T06:32:15,629 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:15,631 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=13, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=test-1691649135102 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:15,634 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=13 2023-08-10T06:32:15,736 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=13 2023-08-10T06:32:15,888 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2023-08-10T06:32:15,938 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=13 2023-08-10T06:32:16,065 INFO [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => 7f5405ff46e3916fa2d5dcf037138bc3, NAME => 'test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='test-1691649135102', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:16,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=13 2023-08-10T06:32:16,480 DEBUG [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(888): Instantiated test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:16,480 DEBUG [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(1668): Closing 7f5405ff46e3916fa2d5dcf037138bc3, disabling compactions & flushes 2023-08-10T06:32:16,480 INFO [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(1690): Closing region test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:16,480 DEBUG [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:16,480 DEBUG [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. after waiting 0 ms 2023-08-10T06:32:16,480 DEBUG [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:16,480 INFO [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(1909): Closed test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:16,480 DEBUG [RegionOpenAndInit-test-1691649135102-pool-0 {}] regionserver.HRegion(1622): Region close journal for 7f5405ff46e3916fa2d5dcf037138bc3: 2023-08-10T06:32:16,482 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=13, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=test-1691649135102 execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:16,482 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3.","families":{"info":[{"qualifier":"regioninfo","vlen":52,"tag":[],"timestamp":"1691649136482"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649136482"}]},"ts":"1691649136482"} 2023-08-10T06:32:16,484 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:16,485 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=13, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=test-1691649135102 execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:16,486 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"test-1691649135102","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649136485"}]},"ts":"1691649136485"} 2023-08-10T06:32:16,487 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=test-1691649135102, state=ENABLING in hbase:meta 2023-08-10T06:32:16,491 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=14, ppid=13, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=test-1691649135102, region=7f5405ff46e3916fa2d5dcf037138bc3, ASSIGN}] 2023-08-10T06:32:16,494 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=14, ppid=13, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=test-1691649135102, region=7f5405ff46e3916fa2d5dcf037138bc3, ASSIGN 2023-08-10T06:32:16,495 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(262): Starting pid=14, ppid=13, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=test-1691649135102, region=7f5405ff46e3916fa2d5dcf037138bc3, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:16,647 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=14 updating hbase:meta row=7f5405ff46e3916fa2d5dcf037138bc3, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:16,650 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=15, ppid=14, state=RUNNABLE; OpenRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:16,742 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=13 2023-08-10T06:32:16,813 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] handler.AssignRegionHandler(130): Open test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:16,813 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7254): Opening region: {ENCODED => 7f5405ff46e3916fa2d5dcf037138bc3, NAME => 'test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:16,814 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:16,814 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table test-1691649135102 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:16,814 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(888): Instantiated test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:16,814 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7296): checking encryption for 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:16,814 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7299): checking classloading for 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:16,815 INFO [StoreOpener-7f5405ff46e3916fa2d5dcf037138bc3-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family f of region 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:16,817 INFO [StoreOpener-7f5405ff46e3916fa2d5dcf037138bc3-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 7f5405ff46e3916fa2d5dcf037138bc3 columnFamilyName f 2023-08-10T06:32:16,817 DEBUG [StoreOpener-7f5405ff46e3916fa2d5dcf037138bc3-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:16,818 INFO [StoreOpener-7f5405ff46e3916fa2d5dcf037138bc3-1 {}] regionserver.HStore(324): Store=7f5405ff46e3916fa2d5dcf037138bc3/f, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:16,819 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:16,819 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:16,823 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1076): writing seq id for 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:16,829 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:16,830 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1093): Opened 7f5405ff46e3916fa2d5dcf037138bc3; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11838863040, jitterRate=0.10258004069328308}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2023-08-10T06:32:16,831 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(992): Region open journal for 7f5405ff46e3916fa2d5dcf037138bc3: 2023-08-10T06:32:16,832 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegionServer(2552): Post open deploy tasks for test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3., pid=15, masterSystemTime=1691649136803 2023-08-10T06:32:16,834 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegionServer(2579): Finished post open deploy task for test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:16,834 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] handler.AssignRegionHandler(158): Opened test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:16,835 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=14 updating hbase:meta row=7f5405ff46e3916fa2d5dcf037138bc3, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:16,840 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=15, resume processing ppid=14 2023-08-10T06:32:16,840 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=15, ppid=14, state=SUCCESS; OpenRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3, server=jenkins-hbase3.apache.org,34991,1691649122119 in 187 msec 2023-08-10T06:32:16,842 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=14, resume processing ppid=13 2023-08-10T06:32:16,842 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=14, ppid=13, state=SUCCESS; TransitRegionStateProcedure table=test-1691649135102, region=7f5405ff46e3916fa2d5dcf037138bc3, ASSIGN in 349 msec 2023-08-10T06:32:16,843 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(81): pid=13, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=test-1691649135102 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:16,843 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"test-1691649135102","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649136843"}]},"ts":"1691649136843"} 2023-08-10T06:32:16,845 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=test-1691649135102, state=ENABLED in hbase:meta 2023-08-10T06:32:16,849 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(81): pid=13, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=test-1691649135102 execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:16,851 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=13, state=SUCCESS; CreateTableProcedure table=test-1691649135102 in 1.2260 sec 2023-08-10T06:32:17,744 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=13 2023-08-10T06:32:17,744 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: CREATE, Table Name: default:test-1691649135102, procId: 13 completed 2023-08-10T06:32:17,744 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x0a1e0a95 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:17,750 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5a787f51, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:17,752 DEBUG [hconnection-0x6ca4081-metaLookup-shared--pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:17,755 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:55476, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:17,764 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HRegion(8210): writing data to region test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. with WAL disabled. Data may be lost in the event of a crash. 2023-08-10T06:32:17,855 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$4(2331): Client=jenkins//172.31.12.81 create 'ns2:test-16916491351021', {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:17,857 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=16, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=ns2:test-16916491351021 2023-08-10T06:32:17,859 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=16, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=ns2:test-16916491351021 execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:17,859 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:17,859 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(708): Client=jenkins//172.31.12.81 procedure request for creating table: namespace: "ns2" qualifier: "test-16916491351021" procId is: 16 2023-08-10T06:32:17,861 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=16, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=ns2:test-16916491351021 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:17,861 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=16 2023-08-10T06:32:17,963 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=16 2023-08-10T06:32:18,165 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=16 2023-08-10T06:32:18,278 INFO [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => c11396d9b009ba78cea809571c6e38e7, NAME => 'ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='ns2:test-16916491351021', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:18,294 DEBUG [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(888): Instantiated ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:18,294 DEBUG [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(1668): Closing c11396d9b009ba78cea809571c6e38e7, disabling compactions & flushes 2023-08-10T06:32:18,294 INFO [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(1690): Closing region ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:32:18,294 DEBUG [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:32:18,294 DEBUG [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. after waiting 0 ms 2023-08-10T06:32:18,294 DEBUG [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:32:18,294 INFO [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(1909): Closed ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:32:18,294 DEBUG [RegionOpenAndInit-ns2:test-16916491351021-pool-0 {}] regionserver.HRegion(1622): Region close journal for c11396d9b009ba78cea809571c6e38e7: 2023-08-10T06:32:18,296 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=16, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=ns2:test-16916491351021 execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:18,296 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7.","families":{"info":[{"qualifier":"regioninfo","vlen":49,"tag":[],"timestamp":"1691649138296"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649138296"}]},"ts":"1691649138296"} 2023-08-10T06:32:18,298 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:18,299 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=16, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=ns2:test-16916491351021 execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:18,300 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"ns2:test-16916491351021","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649138299"}]},"ts":"1691649138299"} 2023-08-10T06:32:18,301 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=ns2:test-16916491351021, state=ENABLING in hbase:meta 2023-08-10T06:32:18,308 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=17, ppid=16, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=ns2:test-16916491351021, region=c11396d9b009ba78cea809571c6e38e7, ASSIGN}] 2023-08-10T06:32:18,311 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=17, ppid=16, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=ns2:test-16916491351021, region=c11396d9b009ba78cea809571c6e38e7, ASSIGN 2023-08-10T06:32:18,312 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(262): Starting pid=17, ppid=16, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=ns2:test-16916491351021, region=c11396d9b009ba78cea809571c6e38e7, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:18,464 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=17 updating hbase:meta row=c11396d9b009ba78cea809571c6e38e7, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:18,467 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=18, ppid=17, state=RUNNABLE; OpenRegionProcedure c11396d9b009ba78cea809571c6e38e7, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:18,467 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=16 2023-08-10T06:32:18,625 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] handler.AssignRegionHandler(130): Open ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:32:18,625 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7254): Opening region: {ENCODED => c11396d9b009ba78cea809571c6e38e7, NAME => 'ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:18,626 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:18,626 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table test-16916491351021 c11396d9b009ba78cea809571c6e38e7 2023-08-10T06:32:18,627 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(888): Instantiated ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:18,627 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7296): checking encryption for c11396d9b009ba78cea809571c6e38e7 2023-08-10T06:32:18,627 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7299): checking classloading for c11396d9b009ba78cea809571c6e38e7 2023-08-10T06:32:18,629 INFO [StoreOpener-c11396d9b009ba78cea809571c6e38e7-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family f of region c11396d9b009ba78cea809571c6e38e7 2023-08-10T06:32:18,631 INFO [StoreOpener-c11396d9b009ba78cea809571c6e38e7-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region c11396d9b009ba78cea809571c6e38e7 columnFamilyName f 2023-08-10T06:32:18,631 DEBUG [StoreOpener-c11396d9b009ba78cea809571c6e38e7-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:18,632 INFO [StoreOpener-c11396d9b009ba78cea809571c6e38e7-1 {}] regionserver.HStore(324): Store=c11396d9b009ba78cea809571c6e38e7/f, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:18,634 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7 2023-08-10T06:32:18,635 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7 2023-08-10T06:32:18,642 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(1076): writing seq id for c11396d9b009ba78cea809571c6e38e7 2023-08-10T06:32:18,657 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:18,658 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(1093): Opened c11396d9b009ba78cea809571c6e38e7; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=9612311840, jitterRate=-0.10478369891643524}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2023-08-10T06:32:18,659 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(992): Region open journal for c11396d9b009ba78cea809571c6e38e7: 2023-08-10T06:32:18,661 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegionServer(2552): Post open deploy tasks for ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7., pid=18, masterSystemTime=1691649138620 2023-08-10T06:32:18,668 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegionServer(2579): Finished post open deploy task for ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:32:18,668 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] handler.AssignRegionHandler(158): Opened ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:32:18,669 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=17 updating hbase:meta row=c11396d9b009ba78cea809571c6e38e7, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:18,678 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=18, resume processing ppid=17 2023-08-10T06:32:18,678 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=18, ppid=17, state=SUCCESS; OpenRegionProcedure c11396d9b009ba78cea809571c6e38e7, server=jenkins-hbase3.apache.org,34991,1691649122119 in 207 msec 2023-08-10T06:32:18,691 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=17, resume processing ppid=16 2023-08-10T06:32:18,691 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=17, ppid=16, state=SUCCESS; TransitRegionStateProcedure table=ns2:test-16916491351021, region=c11396d9b009ba78cea809571c6e38e7, ASSIGN in 370 msec 2023-08-10T06:32:18,691 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=16, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=ns2:test-16916491351021 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:18,692 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"ns2:test-16916491351021","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649138691"}]},"ts":"1691649138691"} 2023-08-10T06:32:18,694 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=ns2:test-16916491351021, state=ENABLED in hbase:meta 2023-08-10T06:32:18,697 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=16, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=ns2:test-16916491351021 execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:18,705 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=16, state=SUCCESS; CreateTableProcedure table=ns2:test-16916491351021 in 842 msec 2023-08-10T06:32:18,969 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=16 2023-08-10T06:32:18,969 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: CREATE, Table Name: ns2:test-16916491351021, procId: 16 completed 2023-08-10T06:32:18,975 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HRegion(8210): writing data to region ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. with WAL disabled. Data may be lost in the event of a crash. 2023-08-10T06:32:19,061 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$4(2331): Client=jenkins//172.31.12.81 create 'ns3:test-16916491351022', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:19,062 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=19, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=ns3:test-16916491351022 2023-08-10T06:32:19,064 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=19, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=ns3:test-16916491351022 execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:19,065 DEBUG [PEWorker-5 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:19,065 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(708): Client=jenkins//172.31.12.81 procedure request for creating table: namespace: "ns3" qualifier: "test-16916491351022" procId is: 19 2023-08-10T06:32:19,066 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=19, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=ns3:test-16916491351022 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:19,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=19 2023-08-10T06:32:19,167 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=19 2023-08-10T06:32:19,308 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_ns2_table_test-16916491351021 2023-08-10T06:32:19,308 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_ns2_table_test-16916491351021 Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:19,309 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_test-1691649135102 2023-08-10T06:32:19,309 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_test-1691649135102 Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:19,369 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=19 2023-08-10T06:32:19,479 INFO [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => 7ecca0e4e0228e4c7a687e00712cb8f1, NAME => 'ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='ns3:test-16916491351022', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:19,671 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=19 2023-08-10T06:32:19,895 DEBUG [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(888): Instantiated ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:19,895 DEBUG [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(1668): Closing 7ecca0e4e0228e4c7a687e00712cb8f1, disabling compactions & flushes 2023-08-10T06:32:19,895 INFO [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(1690): Closing region ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:32:19,895 DEBUG [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:32:19,895 DEBUG [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. after waiting 0 ms 2023-08-10T06:32:19,895 DEBUG [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:32:19,895 INFO [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(1909): Closed ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:32:19,895 DEBUG [RegionOpenAndInit-ns3:test-16916491351022-pool-0 {}] regionserver.HRegion(1622): Region close journal for 7ecca0e4e0228e4c7a687e00712cb8f1: 2023-08-10T06:32:19,897 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=19, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=ns3:test-16916491351022 execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:19,897 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1.","families":{"info":[{"qualifier":"regioninfo","vlen":49,"tag":[],"timestamp":"1691649139897"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649139897"}]},"ts":"1691649139897"} 2023-08-10T06:32:19,899 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:19,901 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=19, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=ns3:test-16916491351022 execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:19,901 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"ns3:test-16916491351022","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649139901"}]},"ts":"1691649139901"} 2023-08-10T06:32:19,903 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=ns3:test-16916491351022, state=ENABLING in hbase:meta 2023-08-10T06:32:19,906 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=20, ppid=19, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=ns3:test-16916491351022, region=7ecca0e4e0228e4c7a687e00712cb8f1, ASSIGN}] 2023-08-10T06:32:19,909 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=20, ppid=19, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=ns3:test-16916491351022, region=7ecca0e4e0228e4c7a687e00712cb8f1, ASSIGN 2023-08-10T06:32:19,909 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(262): Starting pid=20, ppid=19, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=ns3:test-16916491351022, region=7ecca0e4e0228e4c7a687e00712cb8f1, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:20,061 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=20 updating hbase:meta row=7ecca0e4e0228e4c7a687e00712cb8f1, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:20,064 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=21, ppid=20, state=RUNNABLE; OpenRegionProcedure 7ecca0e4e0228e4c7a687e00712cb8f1, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:20,173 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=19 2023-08-10T06:32:20,226 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] handler.AssignRegionHandler(130): Open ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:32:20,226 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(7254): Opening region: {ENCODED => 7ecca0e4e0228e4c7a687e00712cb8f1, NAME => 'ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:20,227 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:20,227 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table test-16916491351022 7ecca0e4e0228e4c7a687e00712cb8f1 2023-08-10T06:32:20,227 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(888): Instantiated ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:20,227 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(7296): checking encryption for 7ecca0e4e0228e4c7a687e00712cb8f1 2023-08-10T06:32:20,227 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(7299): checking classloading for 7ecca0e4e0228e4c7a687e00712cb8f1 2023-08-10T06:32:20,229 INFO [StoreOpener-7ecca0e4e0228e4c7a687e00712cb8f1-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family f of region 7ecca0e4e0228e4c7a687e00712cb8f1 2023-08-10T06:32:20,230 INFO [StoreOpener-7ecca0e4e0228e4c7a687e00712cb8f1-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 7ecca0e4e0228e4c7a687e00712cb8f1 columnFamilyName f 2023-08-10T06:32:20,231 DEBUG [StoreOpener-7ecca0e4e0228e4c7a687e00712cb8f1-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:20,231 INFO [StoreOpener-7ecca0e4e0228e4c7a687e00712cb8f1-1 {}] regionserver.HStore(324): Store=7ecca0e4e0228e4c7a687e00712cb8f1/f, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:20,232 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns3/test-16916491351022/7ecca0e4e0228e4c7a687e00712cb8f1 2023-08-10T06:32:20,233 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns3/test-16916491351022/7ecca0e4e0228e4c7a687e00712cb8f1 2023-08-10T06:32:20,236 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(1076): writing seq id for 7ecca0e4e0228e4c7a687e00712cb8f1 2023-08-10T06:32:20,239 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns3/test-16916491351022/7ecca0e4e0228e4c7a687e00712cb8f1/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:20,240 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(1093): Opened 7ecca0e4e0228e4c7a687e00712cb8f1; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11370770560, jitterRate=0.05898553133010864}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2023-08-10T06:32:20,240 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegion(992): Region open journal for 7ecca0e4e0228e4c7a687e00712cb8f1: 2023-08-10T06:32:20,241 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegionServer(2552): Post open deploy tasks for ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1., pid=21, masterSystemTime=1691649140217 2023-08-10T06:32:20,244 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] regionserver.HRegionServer(2579): Finished post open deploy task for ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:32:20,244 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=21}] handler.AssignRegionHandler(158): Opened ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:32:20,245 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=20 updating hbase:meta row=7ecca0e4e0228e4c7a687e00712cb8f1, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:20,250 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=21, resume processing ppid=20 2023-08-10T06:32:20,250 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=21, ppid=20, state=SUCCESS; OpenRegionProcedure 7ecca0e4e0228e4c7a687e00712cb8f1, server=jenkins-hbase3.apache.org,34991,1691649122119 in 183 msec 2023-08-10T06:32:20,253 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=20, resume processing ppid=19 2023-08-10T06:32:20,253 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=20, ppid=19, state=SUCCESS; TransitRegionStateProcedure table=ns3:test-16916491351022, region=7ecca0e4e0228e4c7a687e00712cb8f1, ASSIGN in 344 msec 2023-08-10T06:32:20,254 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=19, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=ns3:test-16916491351022 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:20,254 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"ns3:test-16916491351022","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649140254"}]},"ts":"1691649140254"} 2023-08-10T06:32:20,256 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=ns3:test-16916491351022, state=ENABLED in hbase:meta 2023-08-10T06:32:20,258 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=19, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=ns3:test-16916491351022 execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:20,260 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=19, state=SUCCESS; CreateTableProcedure table=ns3:test-16916491351022 in 1.1980 sec 2023-08-10T06:32:21,174 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=19 2023-08-10T06:32:21,174 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: CREATE, Table Name: ns3:test-16916491351022, procId: 19 completed 2023-08-10T06:32:21,175 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(3527): Waiting until all regions of table ns3:test-16916491351022 get assigned. Timeout = 60000ms 2023-08-10T06:32:21,177 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2023-08-10T06:32:21,181 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(3581): All regions for table ns3:test-16916491351022 assigned to meta. Checking AM states. 2023-08-10T06:32:21,181 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2023-08-10T06:32:21,182 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(3601): All regions for table ns3:test-16916491351022 assigned. 2023-08-10T06:32:21,183 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$4(2331): Client=jenkins//172.31.12.81 create 'ns4:test-16916491351023', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:21,185 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=22, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=ns4:test-16916491351023 2023-08-10T06:32:21,187 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=22, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=ns4:test-16916491351023 execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:21,187 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:21,187 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(708): Client=jenkins//172.31.12.81 procedure request for creating table: namespace: "ns4" qualifier: "test-16916491351023" procId is: 22 2023-08-10T06:32:21,188 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=22, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=ns4:test-16916491351023 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:21,188 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=22 2023-08-10T06:32:21,290 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=22 2023-08-10T06:32:21,491 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=22 2023-08-10T06:32:21,602 INFO [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => 66486f64c0e50e4c038f3174e657e6c0, NAME => 'ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='ns4:test-16916491351023', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:21,794 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=22 2023-08-10T06:32:22,018 DEBUG [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(888): Instantiated ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:22,018 DEBUG [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(1668): Closing 66486f64c0e50e4c038f3174e657e6c0, disabling compactions & flushes 2023-08-10T06:32:22,018 INFO [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(1690): Closing region ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:32:22,018 DEBUG [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:32:22,018 DEBUG [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. after waiting 0 ms 2023-08-10T06:32:22,018 DEBUG [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:32:22,018 INFO [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(1909): Closed ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:32:22,018 DEBUG [RegionOpenAndInit-ns4:test-16916491351023-pool-0 {}] regionserver.HRegion(1622): Region close journal for 66486f64c0e50e4c038f3174e657e6c0: 2023-08-10T06:32:22,020 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=22, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=ns4:test-16916491351023 execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:22,020 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0.","families":{"info":[{"qualifier":"regioninfo","vlen":49,"tag":[],"timestamp":"1691649142020"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649142020"}]},"ts":"1691649142020"} 2023-08-10T06:32:22,023 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:22,025 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=22, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=ns4:test-16916491351023 execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:22,025 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"ns4:test-16916491351023","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649142025"}]},"ts":"1691649142025"} 2023-08-10T06:32:22,026 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=ns4:test-16916491351023, state=ENABLING in hbase:meta 2023-08-10T06:32:22,029 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=23, ppid=22, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=ns4:test-16916491351023, region=66486f64c0e50e4c038f3174e657e6c0, ASSIGN}] 2023-08-10T06:32:22,032 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=23, ppid=22, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=ns4:test-16916491351023, region=66486f64c0e50e4c038f3174e657e6c0, ASSIGN 2023-08-10T06:32:22,033 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(262): Starting pid=23, ppid=22, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=ns4:test-16916491351023, region=66486f64c0e50e4c038f3174e657e6c0, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:22,185 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=23 updating hbase:meta row=66486f64c0e50e4c038f3174e657e6c0, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:22,187 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=24, ppid=23, state=RUNNABLE; OpenRegionProcedure 66486f64c0e50e4c038f3174e657e6c0, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:22,295 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=22 2023-08-10T06:32:22,318 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2023-08-10T06:32:22,343 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] handler.AssignRegionHandler(130): Open ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:32:22,343 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(7254): Opening region: {ENCODED => 66486f64c0e50e4c038f3174e657e6c0, NAME => 'ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:22,343 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:22,343 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table test-16916491351023 66486f64c0e50e4c038f3174e657e6c0 2023-08-10T06:32:22,344 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(888): Instantiated ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:22,344 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(7296): checking encryption for 66486f64c0e50e4c038f3174e657e6c0 2023-08-10T06:32:22,344 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(7299): checking classloading for 66486f64c0e50e4c038f3174e657e6c0 2023-08-10T06:32:22,346 INFO [StoreOpener-66486f64c0e50e4c038f3174e657e6c0-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family f of region 66486f64c0e50e4c038f3174e657e6c0 2023-08-10T06:32:22,348 INFO [StoreOpener-66486f64c0e50e4c038f3174e657e6c0-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 66486f64c0e50e4c038f3174e657e6c0 columnFamilyName f 2023-08-10T06:32:22,348 DEBUG [StoreOpener-66486f64c0e50e4c038f3174e657e6c0-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:22,349 INFO [StoreOpener-66486f64c0e50e4c038f3174e657e6c0-1 {}] regionserver.HStore(324): Store=66486f64c0e50e4c038f3174e657e6c0/f, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:22,350 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns4/test-16916491351023/66486f64c0e50e4c038f3174e657e6c0 2023-08-10T06:32:22,351 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns4/test-16916491351023/66486f64c0e50e4c038f3174e657e6c0 2023-08-10T06:32:22,362 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(1076): writing seq id for 66486f64c0e50e4c038f3174e657e6c0 2023-08-10T06:32:22,369 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns4/test-16916491351023/66486f64c0e50e4c038f3174e657e6c0/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:22,370 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(1093): Opened 66486f64c0e50e4c038f3174e657e6c0; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11889610880, jitterRate=0.10730630159378052}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2023-08-10T06:32:22,370 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegion(992): Region open journal for 66486f64c0e50e4c038f3174e657e6c0: 2023-08-10T06:32:22,371 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegionServer(2552): Post open deploy tasks for ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0., pid=24, masterSystemTime=1691649142338 2023-08-10T06:32:22,373 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] regionserver.HRegionServer(2579): Finished post open deploy task for ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:32:22,373 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=24}] handler.AssignRegionHandler(158): Opened ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:32:22,374 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=23 updating hbase:meta row=66486f64c0e50e4c038f3174e657e6c0, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:22,378 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=24, resume processing ppid=23 2023-08-10T06:32:22,378 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=24, ppid=23, state=SUCCESS; OpenRegionProcedure 66486f64c0e50e4c038f3174e657e6c0, server=jenkins-hbase3.apache.org,34991,1691649122119 in 190 msec 2023-08-10T06:32:22,381 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=23, resume processing ppid=22 2023-08-10T06:32:22,381 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1414): Finished pid=23, ppid=22, state=SUCCESS; TransitRegionStateProcedure table=ns4:test-16916491351023, region=66486f64c0e50e4c038f3174e657e6c0, ASSIGN in 349 msec 2023-08-10T06:32:22,382 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(81): pid=22, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=ns4:test-16916491351023 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:22,382 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"ns4:test-16916491351023","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649142382"}]},"ts":"1691649142382"} 2023-08-10T06:32:22,384 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=ns4:test-16916491351023, state=ENABLED in hbase:meta 2023-08-10T06:32:22,386 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(81): pid=22, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=ns4:test-16916491351023 execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:22,388 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=22, state=SUCCESS; CreateTableProcedure table=ns4:test-16916491351023 in 1.2030 sec 2023-08-10T06:32:22,423 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'ns4:test-16916491351023' 2023-08-10T06:32:22,423 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'ns2:test-16916491351021' 2023-08-10T06:32:22,424 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'test-1691649135102' 2023-08-10T06:32:22,424 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'ns3:test-16916491351022' 2023-08-10T06:32:23,297 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=22 2023-08-10T06:32:23,297 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: CREATE, Table Name: ns4:test-16916491351023, procId: 22 completed 2023-08-10T06:32:23,297 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(3527): Waiting until all regions of table ns4:test-16916491351023 get assigned. Timeout = 60000ms 2023-08-10T06:32:23,297 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2023-08-10T06:32:23,301 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(3581): All regions for table ns4:test-16916491351023 assigned to meta. Checking AM states. 2023-08-10T06:32:23,301 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2023-08-10T06:32:23,302 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(3601): All regions for table ns4:test-16916491351023 assigned. 2023-08-10T06:32:23,302 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x0a1e0a95 to 127.0.0.1:52413 2023-08-10T06:32:23,302 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:23,366 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.ResourceChecker(147): before: backup.TestRemoteRestore#testFullRestoreRemote Thread=815, OpenFileDescriptor=878, MaxFileDescriptor=60000, SystemLoadAverage=93, ProcessCount=169, AvailableMemoryMB=2357 2023-08-10T06:32:23,366 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.ResourceChecker(130): Thread=815 is superior to 500 2023-08-10T06:32:23,368 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] backup.TestRemoteRestore(70): test remote full backup on a single table 2023-08-10T06:32:23,369 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x2f62e434 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:23,374 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@288bef35, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:23,398 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:23,400 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:43226, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:23,413 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:23,415 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:40878, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:32:23,424 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$17(3366): Client=jenkins//172.31.12.81 creating {NAME => 'backup'} 2023-08-10T06:32:23,425 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=25, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=backup 2023-08-10T06:32:23,430 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=25 2023-08-10T06:32:23,435 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/namespace 2023-08-10T06:32:23,439 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=25, state=SUCCESS; CreateNamespaceProcedure, namespace=backup in 14 msec 2023-08-10T06:32:23,532 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=25 2023-08-10T06:32:23,551 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$4(2331): Client=jenkins//172.31.12.81 create 'backup:system', {NAME => 'meta', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'session', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:23,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=26, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=backup:system 2023-08-10T06:32:23,558 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=26, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=backup:system execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:23,558 DEBUG [PEWorker-5 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:23,558 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(708): Client=jenkins//172.31.12.81 procedure request for creating table: namespace: "backup" qualifier: "system" procId is: 26 2023-08-10T06:32:23,560 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=26, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=backup:system execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:23,560 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=26 2023-08-10T06:32:23,662 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=26 2023-08-10T06:32:23,864 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=26 2023-08-10T06:32:23,982 INFO [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => de2170c3441572f6a0fca17a49c93cc1, NAME => 'backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='backup:system', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'meta', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'session', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:24,166 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=26 2023-08-10T06:32:24,395 DEBUG [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(888): Instantiated backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:24,395 DEBUG [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(1668): Closing de2170c3441572f6a0fca17a49c93cc1, disabling compactions & flushes 2023-08-10T06:32:24,395 INFO [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(1690): Closing region backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:24,395 DEBUG [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:24,395 DEBUG [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. after waiting 0 ms 2023-08-10T06:32:24,395 DEBUG [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:24,395 INFO [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(1909): Closed backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:24,395 DEBUG [RegionOpenAndInit-backup:system-pool-0 {}] regionserver.HRegion(1622): Region close journal for de2170c3441572f6a0fca17a49c93cc1: 2023-08-10T06:32:24,397 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=26, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=backup:system execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:24,398 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1.","families":{"info":[{"qualifier":"regioninfo","vlen":39,"tag":[],"timestamp":"1691649144397"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649144397"}]},"ts":"1691649144397"} 2023-08-10T06:32:24,400 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:24,401 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=26, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=backup:system execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:24,402 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"backup:system","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649144401"}]},"ts":"1691649144401"} 2023-08-10T06:32:24,404 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=backup:system, state=ENABLING in hbase:meta 2023-08-10T06:32:24,407 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=27, ppid=26, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=backup:system, region=de2170c3441572f6a0fca17a49c93cc1, ASSIGN}] 2023-08-10T06:32:24,411 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=27, ppid=26, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=backup:system, region=de2170c3441572f6a0fca17a49c93cc1, ASSIGN 2023-08-10T06:32:24,412 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(262): Starting pid=27, ppid=26, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=backup:system, region=de2170c3441572f6a0fca17a49c93cc1, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:24,564 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=27 updating hbase:meta row=de2170c3441572f6a0fca17a49c93cc1, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:24,567 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=28, ppid=27, state=RUNNABLE; OpenRegionProcedure de2170c3441572f6a0fca17a49c93cc1, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:24,667 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=26 2023-08-10T06:32:24,723 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] handler.AssignRegionHandler(130): Open backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:24,723 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(7254): Opening region: {ENCODED => de2170c3441572f6a0fca17a49c93cc1, NAME => 'backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:24,724 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:24,724 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table system de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:24,724 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(888): Instantiated backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:24,724 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(7296): checking encryption for de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:24,724 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(7299): checking classloading for de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:24,726 INFO [StoreOpener-de2170c3441572f6a0fca17a49c93cc1-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family meta of region de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:24,727 INFO [StoreOpener-de2170c3441572f6a0fca17a49c93cc1-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region de2170c3441572f6a0fca17a49c93cc1 columnFamilyName meta 2023-08-10T06:32:24,727 DEBUG [StoreOpener-de2170c3441572f6a0fca17a49c93cc1-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:24,728 INFO [StoreOpener-de2170c3441572f6a0fca17a49c93cc1-1 {}] regionserver.HStore(324): Store=de2170c3441572f6a0fca17a49c93cc1/meta, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:24,728 INFO [StoreOpener-de2170c3441572f6a0fca17a49c93cc1-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family session of region de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:24,730 INFO [StoreOpener-de2170c3441572f6a0fca17a49c93cc1-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region de2170c3441572f6a0fca17a49c93cc1 columnFamilyName session 2023-08-10T06:32:24,730 DEBUG [StoreOpener-de2170c3441572f6a0fca17a49c93cc1-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:24,731 INFO [StoreOpener-de2170c3441572f6a0fca17a49c93cc1-1 {}] regionserver.HStore(324): Store=de2170c3441572f6a0fca17a49c93cc1/session, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:24,732 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:24,732 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:24,734 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table backup:system descriptor;using region.getMemStoreFlushHeapSize/# of families (64.0 M)) instead. 2023-08-10T06:32:24,735 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(1076): writing seq id for de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:24,737 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:24,740 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(1093): Opened de2170c3441572f6a0fca17a49c93cc1; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11056582560, jitterRate=0.029724493622779846}}}, FlushLargeStoresPolicy{flushSizeLowerBound=67108864} 2023-08-10T06:32:24,740 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegion(992): Region open journal for de2170c3441572f6a0fca17a49c93cc1: 2023-08-10T06:32:24,741 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegionServer(2552): Post open deploy tasks for backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1., pid=28, masterSystemTime=1691649144719 2023-08-10T06:32:24,743 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] regionserver.HRegionServer(2579): Finished post open deploy task for backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:24,743 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=28}] handler.AssignRegionHandler(158): Opened backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:24,744 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=27 updating hbase:meta row=de2170c3441572f6a0fca17a49c93cc1, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:24,748 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=28, resume processing ppid=27 2023-08-10T06:32:24,748 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=28, ppid=27, state=SUCCESS; OpenRegionProcedure de2170c3441572f6a0fca17a49c93cc1, server=jenkins-hbase3.apache.org,34991,1691649122119 in 179 msec 2023-08-10T06:32:24,750 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=27, resume processing ppid=26 2023-08-10T06:32:24,750 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=27, ppid=26, state=SUCCESS; TransitRegionStateProcedure table=backup:system, region=de2170c3441572f6a0fca17a49c93cc1, ASSIGN in 341 msec 2023-08-10T06:32:24,750 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=26, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=backup:system execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:24,751 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"backup:system","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649144750"}]},"ts":"1691649144750"} 2023-08-10T06:32:24,752 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=backup:system, state=ENABLED in hbase:meta 2023-08-10T06:32:24,755 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=26, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=backup:system execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:24,756 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=26, state=SUCCESS; CreateTableProcedure table=backup:system in 1.2040 sec 2023-08-10T06:32:25,669 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=26 2023-08-10T06:32:25,669 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: CREATE, Table Name: backup:system, procId: 26 completed 2023-08-10T06:32:25,685 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$4(2331): Client=jenkins//172.31.12.81 create 'backup:system_bulk', {NAME => 'meta', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'session', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:25,686 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=29, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=backup:system_bulk 2023-08-10T06:32:25,690 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=29, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=backup:system_bulk execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:25,690 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:25,690 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(708): Client=jenkins//172.31.12.81 procedure request for creating table: namespace: "backup" qualifier: "system_bulk" procId is: 29 2023-08-10T06:32:25,691 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=29, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=backup:system_bulk execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:25,692 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=29 2023-08-10T06:32:25,705 INFO [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => 5a53179ddca6d17f3904c3231abeb782, NAME => 'backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='backup:system_bulk', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'meta', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'session', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:25,718 DEBUG [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(888): Instantiated backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:25,718 DEBUG [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(1668): Closing 5a53179ddca6d17f3904c3231abeb782, disabling compactions & flushes 2023-08-10T06:32:25,718 INFO [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(1690): Closing region backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:32:25,718 DEBUG [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:32:25,718 DEBUG [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. after waiting 0 ms 2023-08-10T06:32:25,718 DEBUG [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:32:25,718 INFO [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(1909): Closed backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:32:25,718 DEBUG [RegionOpenAndInit-backup:system_bulk-pool-0 {}] regionserver.HRegion(1622): Region close journal for 5a53179ddca6d17f3904c3231abeb782: 2023-08-10T06:32:25,719 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=29, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=backup:system_bulk execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:25,720 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782.","families":{"info":[{"qualifier":"regioninfo","vlen":44,"tag":[],"timestamp":"1691649145719"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649145719"}]},"ts":"1691649145719"} 2023-08-10T06:32:25,722 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:25,723 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(81): pid=29, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=backup:system_bulk execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:25,723 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"backup:system_bulk","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649145723"}]},"ts":"1691649145723"} 2023-08-10T06:32:25,725 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=backup:system_bulk, state=ENABLING in hbase:meta 2023-08-10T06:32:25,730 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=30, ppid=29, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=backup:system_bulk, region=5a53179ddca6d17f3904c3231abeb782, ASSIGN}] 2023-08-10T06:32:25,732 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=30, ppid=29, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=backup:system_bulk, region=5a53179ddca6d17f3904c3231abeb782, ASSIGN 2023-08-10T06:32:25,734 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(262): Starting pid=30, ppid=29, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=backup:system_bulk, region=5a53179ddca6d17f3904c3231abeb782, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:25,793 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=29 2023-08-10T06:32:25,886 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=30 updating hbase:meta row=5a53179ddca6d17f3904c3231abeb782, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:25,889 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=31, ppid=30, state=RUNNABLE; OpenRegionProcedure 5a53179ddca6d17f3904c3231abeb782, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:25,995 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=29 2023-08-10T06:32:26,046 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] handler.AssignRegionHandler(130): Open backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:32:26,046 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(7254): Opening region: {ENCODED => 5a53179ddca6d17f3904c3231abeb782, NAME => 'backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:26,047 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:26,047 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table system_bulk 5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:32:26,047 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(888): Instantiated backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:26,047 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(7296): checking encryption for 5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:32:26,047 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(7299): checking classloading for 5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:32:26,050 INFO [StoreOpener-5a53179ddca6d17f3904c3231abeb782-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family meta of region 5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:32:26,052 INFO [StoreOpener-5a53179ddca6d17f3904c3231abeb782-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 5a53179ddca6d17f3904c3231abeb782 columnFamilyName meta 2023-08-10T06:32:26,052 DEBUG [StoreOpener-5a53179ddca6d17f3904c3231abeb782-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:26,053 INFO [StoreOpener-5a53179ddca6d17f3904c3231abeb782-1 {}] regionserver.HStore(324): Store=5a53179ddca6d17f3904c3231abeb782/meta, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:26,053 INFO [StoreOpener-5a53179ddca6d17f3904c3231abeb782-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family session of region 5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:32:26,057 INFO [StoreOpener-5a53179ddca6d17f3904c3231abeb782-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 5a53179ddca6d17f3904c3231abeb782 columnFamilyName session 2023-08-10T06:32:26,057 DEBUG [StoreOpener-5a53179ddca6d17f3904c3231abeb782-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:26,058 INFO [StoreOpener-5a53179ddca6d17f3904c3231abeb782-1 {}] regionserver.HStore(324): Store=5a53179ddca6d17f3904c3231abeb782/session, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:26,059 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system_bulk/5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:32:26,059 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system_bulk/5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:32:26,061 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table backup:system_bulk descriptor;using region.getMemStoreFlushHeapSize/# of families (64.0 M)) instead. 2023-08-10T06:32:26,062 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(1076): writing seq id for 5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:32:26,065 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system_bulk/5a53179ddca6d17f3904c3231abeb782/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:26,065 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(1093): Opened 5a53179ddca6d17f3904c3231abeb782; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11703190400, jitterRate=0.08994454145431519}}}, FlushLargeStoresPolicy{flushSizeLowerBound=67108864} 2023-08-10T06:32:26,066 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegion(992): Region open journal for 5a53179ddca6d17f3904c3231abeb782: 2023-08-10T06:32:26,067 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegionServer(2552): Post open deploy tasks for backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782., pid=31, masterSystemTime=1691649146041 2023-08-10T06:32:26,069 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] regionserver.HRegionServer(2579): Finished post open deploy task for backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:32:26,069 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=31}] handler.AssignRegionHandler(158): Opened backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:32:26,070 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=30 updating hbase:meta row=5a53179ddca6d17f3904c3231abeb782, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:26,074 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=31, resume processing ppid=30 2023-08-10T06:32:26,074 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=31, ppid=30, state=SUCCESS; OpenRegionProcedure 5a53179ddca6d17f3904c3231abeb782, server=jenkins-hbase3.apache.org,34991,1691649122119 in 183 msec 2023-08-10T06:32:26,077 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=30, resume processing ppid=29 2023-08-10T06:32:26,077 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=30, ppid=29, state=SUCCESS; TransitRegionStateProcedure table=backup:system_bulk, region=5a53179ddca6d17f3904c3231abeb782, ASSIGN in 345 msec 2023-08-10T06:32:26,078 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=29, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=backup:system_bulk execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:26,078 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"backup:system_bulk","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649146078"}]},"ts":"1691649146078"} 2023-08-10T06:32:26,080 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=backup:system_bulk, state=ENABLED in hbase:meta 2023-08-10T06:32:26,083 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=29, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=backup:system_bulk execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:26,084 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=29, state=SUCCESS; CreateTableProcedure table=backup:system_bulk in 398 msec 2023-08-10T06:32:26,297 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=29 2023-08-10T06:32:26,297 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: CREATE, Table Name: backup:system_bulk, procId: 29 completed 2023-08-10T06:32:26,312 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(610): Start new backup exclusive operation 2023-08-10T06:32:26,359 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1685): Client=jenkins//172.31.12.81 snapshot request for:{ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } 2023-08-10T06:32:26,359 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1691649146359 (current time:1691649146359). 2023-08-10T06:32:26,359 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2023-08-10T06:32:26,360 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snapshot_backup_system VERSION not specified, setting to 2 2023-08-10T06:32:26,360 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x20c762f4 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:26,366 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@218f8676, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:26,371 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:26,372 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:43240, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:26,372 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x20c762f4 to 127.0.0.1:52413 2023-08-10T06:32:26,373 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:26,377 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(770): No existing snapshot, attempting snapshot... 2023-08-10T06:32:26,386 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=32, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } 2023-08-10T06:32:26,388 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(1404): register snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 }, snapshot procedure id = 32 2023-08-10T06:32:26,388 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2023-08-10T06:32:26,389 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=32 2023-08-10T06:32:26,394 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2023-08-10T06:32:26,396 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2023-08-10T06:32:26,490 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=32 2023-08-10T06:32:26,692 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=32 2023-08-10T06:32:26,806 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2023-08-10T06:32:26,809 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=33, ppid=32, state=RUNNABLE; SnapshotRegionProcedure de2170c3441572f6a0fca17a49c93cc1}] 2023-08-10T06:32:26,817 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=33, ppid=32, state=RUNNABLE; SnapshotRegionProcedure de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:26,976 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=34991 {}] regionserver.RSRpcServices(3924): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=33 2023-08-10T06:32:26,977 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:26,980 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.HRegion(2819): Flushing de2170c3441572f6a0fca17a49c93cc1 2/2 column families, dataSize=45 B heapSize=632 B 2023-08-10T06:32:26,993 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=32 2023-08-10T06:32:27,057 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/a7aa6b78efc846bebb42d595b96dfe44 is 49, key is activesession:/session:c/1691649146345/Put/seqid=0 2023-08-10T06:32:27,479 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=45 B at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/a7aa6b78efc846bebb42d595b96dfe44 2023-08-10T06:32:27,494 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=32 2023-08-10T06:32:27,541 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: a7aa6b78efc846bebb42d595b96dfe44 2023-08-10T06:32:27,544 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/a7aa6b78efc846bebb42d595b96dfe44 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/a7aa6b78efc846bebb42d595b96dfe44 2023-08-10T06:32:27,551 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/a7aa6b78efc846bebb42d595b96dfe44, entries=1, sequenceid=5, filesize=4.9 K 2023-08-10T06:32:27,555 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.HRegion(3022): Finished flush of dataSize ~45 B/45, heapSize ~360 B/360, currentSize=0 B/0 for de2170c3441572f6a0fca17a49c93cc1 in 575ms, sequenceid=5, compaction requested=false 2023-08-10T06:32:27,555 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'backup:system' 2023-08-10T06:32:27,557 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.HRegion(2520): Flush status journal for de2170c3441572f6a0fca17a49c93cc1: 2023-08-10T06:32:27,557 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.SnapshotRegionCallable(75): Snapshotting region backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. for snapshot_backup_system completed. 2023-08-10T06:32:27,558 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] snapshot.SnapshotManifest(240): Storing 'backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1.' region-info for snapshot=snapshot_backup_system 2023-08-10T06:32:27,564 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] snapshot.SnapshotManifest(245): Creating references for hfiles 2023-08-10T06:32:27,569 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] snapshot.SnapshotManifest(255): Adding snapshot references for [] hfiles 2023-08-10T06:32:27,569 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] snapshot.SnapshotManifest(255): Adding snapshot references for [hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/a7aa6b78efc846bebb42d595b96dfe44] hfiles 2023-08-10T06:32:27,569 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] snapshot.SnapshotManifest(264): Adding reference for file (1/1): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/a7aa6b78efc846bebb42d595b96dfe44 for snapshot=snapshot_backup_system 2023-08-10T06:32:27,875 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2023-08-10T06:32:27,984 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:27,988 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=33}] regionserver.RemoteProcedureResultReporter(61): Successfully complete execution of pid=33 2023-08-10T06:32:27,991 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster(3994): Remote procedure done, pid=33 2023-08-10T06:32:27,991 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure.SnapshotRegionProcedure(130): finish snapshot snapshot_backup_system on region de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:27,993 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=33, ppid=32, state=RUNNABLE; SnapshotRegionProcedure de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:28,001 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=33, resume processing ppid=32 2023-08-10T06:32:28,001 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2023-08-10T06:32:28,001 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1414): Finished pid=33, ppid=32, state=SUCCESS; SnapshotRegionProcedure de2170c3441572f6a0fca17a49c93cc1 in 1.1850 sec 2023-08-10T06:32:28,003 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2023-08-10T06:32:28,006 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2023-08-10T06:32:28,006 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(489): Convert to Single Snapshot Manifest for snapshot_backup_system 2023-08-10T06:32:28,009 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(126): No regions under directory:hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/.tmp/snapshot_backup_system 2023-08-10T06:32:28,425 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'backup:system_bulk' 2023-08-10T06:32:28,435 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2023-08-10T06:32:28,447 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2023-08-10T06:32:28,448 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/.tmp/snapshot_backup_system to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/snapshot_backup_system 2023-08-10T06:32:28,495 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=32 2023-08-10T06:32:29,279 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=32, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2023-08-10T06:32:29,279 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1410): unregister snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 }, snapshot procedure id = 32 2023-08-10T06:32:29,281 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=32, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=32, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } in 2.8980 sec 2023-08-10T06:32:29,308 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_backup_table_system 2023-08-10T06:32:29,308 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_backup_table_system Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:29,308 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_backup_table_system_bulk 2023-08-10T06:32:29,308 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_backup_table_system_bulk Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:29,309 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_ns4_table_test-16916491351023 2023-08-10T06:32:29,309 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_ns4_table_test-16916491351023 Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:29,309 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_ns3_table_test-16916491351022 2023-08-10T06:32:29,309 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_ns3_table_test-16916491351022 Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:30,497 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=32 2023-08-10T06:32:30,497 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: SNAPSHOT, Table Name: backup:system, procId: 32 completed 2023-08-10T06:32:30,498 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(115): Backup backup_1691649143378 started at 1691649150497. 2023-08-10T06:32:30,520 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(119): Backup session backup_1691649143378 has been started. 2023-08-10T06:32:30,523 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(150): Execute roll log procedure for full backup ... 2023-08-10T06:32:30,534 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(946): Client=jenkins//172.31.12.81 procedure request for: rolllog-proc 2023-08-10T06:32:30,542 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure.ProcedureCoordinator(165): Submitting procedure rolllog 2023-08-10T06:32:30,542 INFO [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(191): Starting procedure 'rolllog' 2023-08-10T06:32:30,542 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] errorhandling.TimeoutExceptionInjector(107): Scheduling process timer to run in: 180000 ms 2023-08-10T06:32:30,543 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(199): Procedure 'rolllog' starting 'acquire' 2023-08-10T06:32:30,543 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(241): Starting procedure 'rolllog', kicking off acquire phase on members. 2023-08-10T06:32:30,544 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,544 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureCoordinator(92): Creating acquire znode:/1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:30,546 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/acquired 2023-08-10T06:32:30,546 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureCoordinator(100): Watching for acquire node:/1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,546 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(104): Received procedure start children changed event: /1/rolllog-proc/acquired 2023-08-10T06:32:30,546 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/1/rolllog-proc/acquired' 2023-08-10T06:32:30,546 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,546 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(203): Waiting for all members to 'acquire' 2023-08-10T06:32:30,546 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(186): Found procedure znode: /1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:30,547 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,547 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(212): start proc data length is 35 2023-08-10T06:32:30,547 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(214): Found data for znode:/1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:30,547 INFO [zk-event-processor-pool-0 {}] regionserver.LogRollRegionServerProcedureManager(126): Attempting to run a roll log procedure for backup. 2023-08-10T06:32:30,549 INFO [zk-event-processor-pool-0 {}] regionserver.LogRollBackupSubprocedure(56): Constructing a LogRollBackupSubprocedure. 2023-08-10T06:32:30,549 DEBUG [zk-event-processor-pool-0 {}] procedure.ProcedureMember(140): Submitting new Subprocedure:rolllog 2023-08-10T06:32:30,550 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(151): Starting subprocedure 'rolllog' with timeout 60000ms 2023-08-10T06:32:30,550 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] errorhandling.TimeoutExceptionInjector(107): Scheduling process timer to run in: 60000 ms 2023-08-10T06:32:30,551 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(159): Subprocedure 'rolllog' starting 'acquire' stage 2023-08-10T06:32:30,551 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(161): Subprocedure 'rolllog' locally acquired 2023-08-10T06:32:30,551 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.ZKProcedureMemberRpcs(242): Member: 'jenkins-hbase3.apache.org,34991,1691649122119' joining acquired barrier for procedure (rolllog) in zk 2023-08-10T06:32:30,552 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.ZKProcedureMemberRpcs(250): Watch for global barrier reached:/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:30,552 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,552 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureCoordinator$1(194): Node created: /1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,552 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(225): Current zk system: 2023-08-10T06:32:30,552 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(227): |-/1/rolllog-proc 2023-08-10T06:32:30,553 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/reached/rolllog 2023-08-10T06:32:30,553 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(166): Subprocedure 'rolllog' coordinator notified of 'acquire', waiting on 'reached' or 'abort' from coordinator 2023-08-10T06:32:30,553 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-abort 2023-08-10T06:32:30,553 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-acquired 2023-08-10T06:32:30,553 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:30,554 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,554 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-reached 2023-08-10T06:32:30,554 DEBUG [zk-event-processor-pool-0 {}] procedure.Procedure(291): member: 'jenkins-hbase3.apache.org,34991,1691649122119' joining acquired barrier for procedure 'rolllog' on coordinator 2023-08-10T06:32:30,554 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(207): Procedure 'rolllog' starting 'in-barrier' execution. 2023-08-10T06:32:30,554 DEBUG [zk-event-processor-pool-0 {}] procedure.Procedure(300): Waiting on: java.util.concurrent.CountDownLatch@59143346[Count = 0] remaining members to acquire global barrier 2023-08-10T06:32:30,554 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureCoordinator(116): Creating reached barrier zk node:/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:30,556 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:30,556 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(77): Received created event:/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:30,556 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(128): Received reached global barrier:/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:30,556 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(180): Subprocedure 'rolllog' received 'reached' from coordinator. 2023-08-10T06:32:30,556 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,556 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(211): Waiting for all members to 'release' 2023-08-10T06:32:30,558 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] regionserver.LogRollBackupSubprocedurePool(84): Waiting for backup procedure to finish. 2023-08-10T06:32:30,558 DEBUG [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] regionserver.LogRollBackupSubprocedure$RSRollLogTask(75): DRPC started: jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,558 INFO [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] regionserver.LogRollBackupSubprocedure$RSRollLogTask(92): Trying to roll log in backup subprocedure, current log number: 1691649124149 highest: 1691649124149 on jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,558 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractWALRoller(197): WAL AsyncFSWAL jenkins-hbase3.apache.org%2C34991%2C1691649122119:(num 1691649124149) roll requested 2023-08-10T06:32:30,569 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649150558, exclude list is [], retry=0 2023-08-10T06:32:30,578 DEBUG [RS-EventLoopGroup-3-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK] 2023-08-10T06:32:30,582 INFO [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(828): Rolled WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649124149 with entries=19, filesize=3.99 KB; new WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649150558 2023-08-10T06:32:30,584 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK]] 2023-08-10T06:32:30,584 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(742): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649124149 is not closed yet, will try archiving it next time 2023-08-10T06:32:30,584 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractWALRoller(197): WAL AsyncFSWAL jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta:.meta(num 1691649125053) roll requested 2023-08-10T06:32:30,597 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649150584.meta, exclude list is [], retry=0 2023-08-10T06:32:30,602 DEBUG [RS-EventLoopGroup-3-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK] 2023-08-10T06:32:30,604 INFO [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(828): Rolled WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649125053.meta with entries=36, filesize=11.13 KB; new WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649150584.meta 2023-08-10T06:32:30,604 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK]] 2023-08-10T06:32:30,604 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(742): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649125053.meta is not closed yet, will try archiving it next time 2023-08-10T06:32:30,619 DEBUG [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] regionserver.LogRollBackupSubprocedure$RSRollLogTask(99): log roll took 61 2023-08-10T06:32:30,619 INFO [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] regionserver.LogRollBackupSubprocedure$RSRollLogTask(100): After roll log in backup subprocedure, current log number: 1691649150558 on jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,619 DEBUG [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:30,621 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:41910, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=MasterService 2023-08-10T06:32:30,644 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(182): Subprocedure 'rolllog' locally completed 2023-08-10T06:32:30,644 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.ZKProcedureMemberRpcs(267): Marking procedure 'rolllog' completed for member 'jenkins-hbase3.apache.org,34991,1691649122119' in zk 2023-08-10T06:32:30,645 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,645 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(187): Subprocedure 'rolllog' has notified controller of completion 2023-08-10T06:32:30,646 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] errorhandling.TimeoutExceptionInjector(87): Marking timer as complete - no error notifications will be received for this timer. 2023-08-10T06:32:30,646 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureCoordinator$1(194): Node created: /1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,646 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(212): Subprocedure 'rolllog' completed. 2023-08-10T06:32:30,647 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(225): Current zk system: 2023-08-10T06:32:30,648 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(227): |-/1/rolllog-proc 2023-08-10T06:32:30,648 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-abort 2023-08-10T06:32:30,648 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-acquired 2023-08-10T06:32:30,649 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:30,649 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,650 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-reached 2023-08-10T06:32:30,650 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:30,650 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,651 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureCoordinator$1(218): Finished data from procedure 'rolllog' member 'jenkins-hbase3.apache.org,34991,1691649122119': 2023-08-10T06:32:30,651 DEBUG [zk-event-processor-pool-0 {}] procedure.Procedure(321): Member: 'jenkins-hbase3.apache.org,34991,1691649122119' released barrier for procedure'rolllog', counting down latch. Waiting for 0 more 2023-08-10T06:32:30,651 INFO [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(216): Procedure 'rolllog' execution completed 2023-08-10T06:32:30,651 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(225): Running finish phase. 2023-08-10T06:32:30,651 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(275): Finished coordinator procedure - removing self from list of running procedures 2023-08-10T06:32:30,652 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureCoordinator(162): Attempting to clean out zk node for op:rolllog 2023-08-10T06:32:30,652 INFO [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureUtil(265): Clearing all znodes for procedure rolllogincluding nodes /1/rolllog-proc/acquired /1/rolllog-proc/reached /1/rolllog-proc/abort 2023-08-10T06:32:30,653 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,653 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,653 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/abort 2023-08-10T06:32:30,653 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureCoordinator$1(194): Node created: /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,653 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(77): Received created event:/1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,653 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(225): Current zk system: 2023-08-10T06:32:30,653 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(227): |-/1/rolllog-proc 2023-08-10T06:32:30,653 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(316): Aborting procedure member for znode /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,654 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-abort 2023-08-10T06:32:30,654 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(107): Received procedure abort children changed event: /1/rolllog-proc/abort 2023-08-10T06:32:30,654 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/1/rolllog-proc/abort' 2023-08-10T06:32:30,654 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(111): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on existing znode=/1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,654 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:30,654 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(316): Aborting procedure member for znode /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,655 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-acquired 2023-08-10T06:32:30,655 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:30,655 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,656 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(111): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on existing znode=/1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,656 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-reached 2023-08-10T06:32:30,658 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:30,658 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/acquired 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/abort 2023-08-10T06:32:30,666 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(104): Received procedure start children changed event: /1/rolllog-proc/acquired 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:30,666 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/1/rolllog-proc/acquired' 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:30,666 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/abort/rolllog 2023-08-10T06:32:30,666 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] errorhandling.TimeoutExceptionInjector(87): Marking timer as complete - no error notifications will be received for this timer. 2023-08-10T06:32:30,666 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.LogRollMasterProcedureManager(143): Done waiting - exec procedure for rolllog 2023-08-10T06:32:30,668 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.LogRollMasterProcedureManager(144): Distributed roll log procedure is successful! 2023-08-10T06:32:30,668 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(107): Received procedure abort children changed event: /1/rolllog-proc/abort 2023-08-10T06:32:30,668 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/1/rolllog-proc/abort' 2023-08-10T06:32:30,670 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin(2751): Waiting a max of 300000 ms for procedure 'rolllog-proc : rolllog'' to complete. (max 6666 ms per retry) 2023-08-10T06:32:30,670 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin(2759): (#1) Sleeping: 100ms while waiting for procedure completion. 2023-08-10T06:32:30,770 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin(2765): Getting current status of procedure from master... 2023-08-10T06:32:30,775 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1247): Checking to see if procedure from request:rolllog-proc is done 2023-08-10T06:32:30,779 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1685): Client=jenkins//172.31.12.81 snapshot request for:{ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } 2023-08-10T06:32:30,779 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1691649150779 (current time:1691649150779). 2023-08-10T06:32:30,779 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2023-08-10T06:32:30,779 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snapshot_1691649150779_default_test-1691649135102 VERSION not specified, setting to 2 2023-08-10T06:32:30,780 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x54e279fe to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:30,786 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2ec64e68, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:30,788 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:30,789 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:46600, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:30,789 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x54e279fe to 127.0.0.1:52413 2023-08-10T06:32:30,790 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:30,790 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(770): No existing snapshot, attempting snapshot... 2023-08-10T06:32:30,791 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=34, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } 2023-08-10T06:32:30,792 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(1404): register snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 }, snapshot procedure id = 34 2023-08-10T06:32:30,793 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2023-08-10T06:32:30,793 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=34 2023-08-10T06:32:30,794 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2023-08-10T06:32:30,796 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2023-08-10T06:32:30,895 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=34 2023-08-10T06:32:31,096 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=34 2023-08-10T06:32:31,210 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2023-08-10T06:32:31,210 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=35, ppid=34, state=RUNNABLE; SnapshotRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3}] 2023-08-10T06:32:31,212 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=35, ppid=34, state=RUNNABLE; SnapshotRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:31,363 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=34991 {}] regionserver.RSRpcServices(3924): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=35 2023-08-10T06:32:31,364 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:31,366 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.HRegion(2819): Flushing 7f5405ff46e3916fa2d5dcf037138bc3 1/1 column families, dataSize=3.17 KB heapSize=11 KB 2023-08-10T06:32:31,388 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/.tmp/f/317474dc8ece4aa0a1b9b2a97a3c57de is 37, key is row10/f:q1/1691649137773/Put/seqid=0 2023-08-10T06:32:31,397 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=34 2023-08-10T06:32:31,795 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=3.17 KB at sequenceid=103 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/.tmp/f/317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:31,804 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:31,805 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/.tmp/f/317474dc8ece4aa0a1b9b2a97a3c57de as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:31,812 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de, entries=99, sequenceid=103, filesize=8.2 K 2023-08-10T06:32:31,814 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.HRegion(3022): Finished flush of dataSize ~3.17 KB/3247, heapSize ~10.98 KB/11248, currentSize=0 B/0 for 7f5405ff46e3916fa2d5dcf037138bc3 in 449ms, sequenceid=103, compaction requested=false 2023-08-10T06:32:31,815 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.HRegion(2520): Flush status journal for 7f5405ff46e3916fa2d5dcf037138bc3: 2023-08-10T06:32:31,815 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.SnapshotRegionCallable(75): Snapshotting region test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. for snapshot_1691649150779_default_test-1691649135102 completed. 2023-08-10T06:32:31,815 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] snapshot.SnapshotManifest(240): Storing 'test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3.' region-info for snapshot=snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:31,815 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] snapshot.SnapshotManifest(245): Creating references for hfiles 2023-08-10T06:32:31,815 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] snapshot.SnapshotManifest(255): Adding snapshot references for [hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de] hfiles 2023-08-10T06:32:31,815 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] snapshot.SnapshotManifest(264): Adding reference for file (1/1): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de for snapshot=snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:31,898 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=34 2023-08-10T06:32:32,222 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:32,222 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=35}] regionserver.RemoteProcedureResultReporter(61): Successfully complete execution of pid=35 2023-08-10T06:32:32,222 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster(3994): Remote procedure done, pid=35 2023-08-10T06:32:32,222 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure.SnapshotRegionProcedure(130): finish snapshot snapshot_1691649150779_default_test-1691649135102 on region 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:32,224 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=35, ppid=34, state=RUNNABLE; SnapshotRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:32,227 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=35, resume processing ppid=34 2023-08-10T06:32:32,227 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2023-08-10T06:32:32,227 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=35, ppid=34, state=SUCCESS; SnapshotRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3 in 1.0150 sec 2023-08-10T06:32:32,228 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2023-08-10T06:32:32,229 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2023-08-10T06:32:32,229 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(489): Convert to Single Snapshot Manifest for snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:32,230 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(126): No regions under directory:hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/.tmp/snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:32,643 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2023-08-10T06:32:32,650 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2023-08-10T06:32:32,650 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/.tmp/snapshot_1691649150779_default_test-1691649135102 to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:32,899 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=34 2023-08-10T06:32:33,471 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=34, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2023-08-10T06:32:33,471 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1410): unregister snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 }, snapshot procedure id = 34 2023-08-10T06:32:33,473 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1414): Finished pid=34, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=34, snapshot={ ss=snapshot_1691649150779_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } in 2.6810 sec 2023-08-10T06:32:34,624 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2023-08-10T06:32:34,626 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:41920, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2023-08-10T06:32:34,901 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=34 2023-08-10T06:32:34,901 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: SNAPSHOT, Table Name: default:test-1691649135102, procId: 34 completed 2023-08-10T06:32:34,901 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(171): snapshot copy for backup_1691649143378 2023-08-10T06:32:34,901 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(68): Snapshot copy is starting. 2023-08-10T06:32:34,903 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(80): There are 1 snapshots to be copied. 2023-08-10T06:32:34,905 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(105): Setting snapshot copy job name to : Full-Backup_backup_1691649143378_test-1691649135102 2023-08-10T06:32:34,905 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(109): Copy snapshot snapshot_1691649150779_default_test-1691649135102 to hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/ 2023-08-10T06:32:34,915 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.MapReduceBackupCopyJob(373): Doing SNAPSHOT_COPY 2023-08-10T06:32:34,944 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1003): inputFs=hdfs://localhost:33549, inputRoot=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:34,944 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1004): outputFs=DFS[DFSClient[clientName=DFSClient_hb_m_jenkins-hbase3.apache.org,42279,1691649119143_-380523106_16, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/.hbase-snapshot/.tmp/snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:34,944 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1009): Verify snapshot source, inputFs=hdfs://localhost:33549, inputRoot=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3, snapshotDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/snapshot_1691649150779_default_test-1691649135102. 2023-08-10T06:32:34,957 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1068): Copy Snapshot Manifest from hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/snapshot_1691649150779_default_test-1691649135102 to hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/.hbase-snapshot/.tmp/snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:34,991 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:34,992 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:34,992 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:34,993 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,980 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop-4000583040134248228.jar 2023-08-10T06:32:35,980 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,981 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,981 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,982 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,982 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,982 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,984 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,985 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,985 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:35,986 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.4/hbase-shaded-miscellaneous-4.1.4.jar 2023-08-10T06:32:35,987 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.4/hbase-shaded-gson-4.1.4.jar 2023-08-10T06:32:35,987 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.4/hbase-shaded-protobuf-4.1.4.jar 2023-08-10T06:32:35,987 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.4/hbase-shaded-netty-4.1.4.jar 2023-08-10T06:32:35,988 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.4/hbase-unsafe-4.1.4.jar 2023-08-10T06:32:35,988 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.5.7/zookeeper-3.5.7.jar 2023-08-10T06:32:35,989 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2023-08-10T06:32:35,989 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2023-08-10T06:32:35,990 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2023-08-10T06:32:35,990 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2023-08-10T06:32:35,991 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2023-08-10T06:32:35,991 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2023-08-10T06:32:35,995 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/2.10.2/hadoop-common-2.10.2.jar 2023-08-10T06:32:35,996 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/2.10.2/hadoop-common-2.10.2.jar 2023-08-10T06:32:35,996 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.10.2/hadoop-mapreduce-client-core-2.10.2.jar 2023-08-10T06:32:35,996 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/2.10.2/hadoop-common-2.10.2.jar 2023-08-10T06:32:35,997 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/2.10.2/hadoop-common-2.10.2.jar 2023-08-10T06:32:35,997 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.10.2/hadoop-mapreduce-client-core-2.10.2.jar 2023-08-10T06:32:35,998 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.10.2/hadoop-mapreduce-client-core-2.10.2.jar 2023-08-10T06:32:36,056 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.JobResourceUploader(142): Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this. 2023-08-10T06:32:36,080 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(581): Loading Snapshot 'snapshot_1691649150779_default_test-1691649135102' hfile list 2023-08-10T06:32:36,089 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(674): export split=0 size=8.2 K 2023-08-10T06:32:36,303 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] fs.FileUtil(888): Command 'ln -s /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.local.dir/1691649156249/libjars /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/*' failed 1 with: ln: failed to create symbolic link '/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/*': No such file or directory 2023-08-10T06:32:36,303 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] mapred.LocalDistributedCacheManager(202): Failed to create symlink: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.local.dir/1691649156249/libjars <- /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/* 2023-08-10T06:32:36,477 INFO [LocalJobRunner Map Task Executor #0 {}] snapshot.ExportSnapshot$ExportMapper(228): Using bufferSize=128 M 2023-08-10T06:32:36,509 INFO [LocalJobRunner Map Task Executor #0 {}] snapshot.ExportSnapshot$ExportMapper(454): copy completed for input=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de output=hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:36,509 INFO [LocalJobRunner Map Task Executor #0 {}] snapshot.ExportSnapshot$ExportMapper(456): size=8396 (8.2 K) time=0sec 8.007M/sec 2023-08-10T06:32:37,370 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1125): Finalize the Snapshot Export 2023-08-10T06:32:37,371 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1136): Verify snapshot integrity 2023-08-10T06:32:37,379 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1140): Export Completed: snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:37,379 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(121): Snapshot copy snapshot_1691649150779_default_test-1691649135102 finished. 2023-08-10T06:32:37,379 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(1014): test-1691649135102 2023-08-10T06:32:37,448 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManager(277): Getting the direct ancestors of the current backup backup_1691649143378 2023-08-10T06:32:37,448 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManager(283): Current backup is a full backup, no direct ancestor for it. 2023-08-10T06:32:37,859 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManifest(484): Manifest file stored to hdfs://localhost:44455/backupUT/backup_1691649143378/.backup.manifest 2023-08-10T06:32:37,860 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(358): Backup backup_1691649143378 finished: type=FULL,tablelist=test-1691649135102,targetRootDir=hdfs://localhost:44455/backupUT,startts=1691649150497,completets=1691649157446,bytescopied=0 2023-08-10T06:32:37,860 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(138): Trying to delete snapshot for full backup. 2023-08-10T06:32:37,860 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(143): Trying to delete snapshot: snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:37,870 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(759): Client=jenkins//172.31.12.81 delete name: "snapshot_1691649150779_default_test-1691649135102" 2023-08-10T06:32:37,873 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(378): Deleting snapshot: snapshot_1691649150779_default_test-1691649135102 2023-08-10T06:32:37,875 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(148): Deleting the snapshot snapshot_1691649150779_default_test-1691649135102 for backup backup_1691649143378 succeeded. 2023-08-10T06:32:37,876 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(1552): Deleting snapshot_backup_system from the system 2023-08-10T06:32:37,887 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(759): Client=jenkins//172.31.12.81 delete name: "snapshot_backup_system" 2023-08-10T06:32:37,889 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(378): Deleting snapshot: snapshot_backup_system 2023-08-10T06:32:37,890 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(1557): Done deleting backup system table snapshot 2023-08-10T06:32:37,897 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(637): Finish backup exclusive operation 2023-08-10T06:32:37,914 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(376): Backup backup_1691649143378 completed. 2023-08-10T06:32:37,914 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:32:37,914 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x2f62e434 to 127.0.0.1:52413 2023-08-10T06:32:37,914 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:37,914 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] backup.TestRemoteRestore(73): backup complete 2023-08-10T06:32:37,918 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.RestoreTablesClient(101): HBase table table1 does not exist. It will be created during restore process 2023-08-10T06:32:37,918 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManifest(413): Loading manifest from: hdfs://localhost:44455/backupUT/backup_1691649143378/.backup.manifest 2023-08-10T06:32:37,921 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManifest(443): Loaded manifest instance from manifest file: /backupUT/backup_1691649143378/.backup.manifest 2023-08-10T06:32:37,928 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManifest(413): Loading manifest from: hdfs://localhost:44455/backupUT/backup_1691649143378/.backup.manifest 2023-08-10T06:32:37,930 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManifest(443): Loaded manifest instance from manifest file: /backupUT/backup_1691649143378/.backup.manifest 2023-08-10T06:32:37,930 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.RestoreTablesClient(148): Restoring 'test-1691649135102' to 'table1' from full backup image hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102 2023-08-10T06:32:37,940 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] util.RestoreTool(487): Creating target table 'table1' 2023-08-10T06:32:37,941 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] util.RestoreTool(391): Parsing region dir: hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:37,941 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] util.RestoreTool(421): Parsing family dir [hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f in region [hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3] 2023-08-10T06:32:37,946 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] util.RestoreTool(442): Trying to figure out region boundaries hfile=hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de first=row0 last=row98 2023-08-10T06:32:37,946 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:37,953 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$4(2331): Client=jenkins//172.31.12.81 create 'table1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:37,955 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=36, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=table1 2023-08-10T06:32:37,957 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(81): pid=36, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:37,957 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(708): Client=jenkins//172.31.12.81 procedure request for creating table: namespace: "default" qualifier: "table1" procId is: 36 2023-08-10T06:32:37,958 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(81): pid=36, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:37,958 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=36 2023-08-10T06:32:38,059 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=36 2023-08-10T06:32:38,261 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=36 2023-08-10T06:32:38,371 INFO [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => 95cd117c0618eb0283c218ddc0657261, NAME => 'table1,,1691649157953.95cd117c0618eb0283c218ddc0657261.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='table1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:38,562 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=36 2023-08-10T06:32:38,777 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(888): Instantiated table1,,1691649157953.95cd117c0618eb0283c218ddc0657261.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:38,777 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1668): Closing 95cd117c0618eb0283c218ddc0657261, disabling compactions & flushes 2023-08-10T06:32:38,778 INFO [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1690): Closing region table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:38,778 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:38,778 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. after waiting 0 ms 2023-08-10T06:32:38,778 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:38,778 INFO [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1909): Closed table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:38,778 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1622): Region close journal for 95cd117c0618eb0283c218ddc0657261: 2023-08-10T06:32:38,779 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(81): pid=36, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:38,779 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"table1,,1691649157953.95cd117c0618eb0283c218ddc0657261.","families":{"info":[{"qualifier":"regioninfo","vlen":40,"tag":[],"timestamp":"1691649158779"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649158779"}]},"ts":"1691649158779"} 2023-08-10T06:32:38,782 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:38,783 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(81): pid=36, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:38,783 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649158783"}]},"ts":"1691649158783"} 2023-08-10T06:32:38,785 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=table1, state=ENABLING in hbase:meta 2023-08-10T06:32:38,797 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=37, ppid=36, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=table1, region=95cd117c0618eb0283c218ddc0657261, ASSIGN}] 2023-08-10T06:32:38,799 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=37, ppid=36, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=table1, region=95cd117c0618eb0283c218ddc0657261, ASSIGN 2023-08-10T06:32:38,800 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(262): Starting pid=37, ppid=36, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=table1, region=95cd117c0618eb0283c218ddc0657261, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:38,952 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=37 updating hbase:meta row=95cd117c0618eb0283c218ddc0657261, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:38,954 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=38, ppid=37, state=RUNNABLE; OpenRegionProcedure 95cd117c0618eb0283c218ddc0657261, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:39,063 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=36 2023-08-10T06:32:39,108 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] handler.AssignRegionHandler(130): Open table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:39,109 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(7254): Opening region: {ENCODED => 95cd117c0618eb0283c218ddc0657261, NAME => 'table1,,1691649157953.95cd117c0618eb0283c218ddc0657261.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:39,109 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:39,109 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table table1 95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:39,109 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(888): Instantiated table1,,1691649157953.95cd117c0618eb0283c218ddc0657261.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:39,109 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(7296): checking encryption for 95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:39,109 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(7299): checking classloading for 95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:39,111 INFO [StoreOpener-95cd117c0618eb0283c218ddc0657261-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family f of region 95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:39,112 INFO [StoreOpener-95cd117c0618eb0283c218ddc0657261-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 95cd117c0618eb0283c218ddc0657261 columnFamilyName f 2023-08-10T06:32:39,112 DEBUG [StoreOpener-95cd117c0618eb0283c218ddc0657261-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:39,113 INFO [StoreOpener-95cd117c0618eb0283c218ddc0657261-1 {}] regionserver.HStore(324): Store=95cd117c0618eb0283c218ddc0657261/f, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:39,113 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:39,114 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:39,116 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(1076): writing seq id for 95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:39,119 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:39,119 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(1093): Opened 95cd117c0618eb0283c218ddc0657261; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=10785119520, jitterRate=0.004442527890205383}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2023-08-10T06:32:39,120 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegion(992): Region open journal for 95cd117c0618eb0283c218ddc0657261: 2023-08-10T06:32:39,121 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegionServer(2552): Post open deploy tasks for table1,,1691649157953.95cd117c0618eb0283c218ddc0657261., pid=38, masterSystemTime=1691649159105 2023-08-10T06:32:39,122 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] regionserver.HRegionServer(2579): Finished post open deploy task for table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:39,122 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=38}] handler.AssignRegionHandler(158): Opened table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:39,123 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=37 updating hbase:meta row=95cd117c0618eb0283c218ddc0657261, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:39,126 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=38, resume processing ppid=37 2023-08-10T06:32:39,126 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1414): Finished pid=38, ppid=37, state=SUCCESS; OpenRegionProcedure 95cd117c0618eb0283c218ddc0657261, server=jenkins-hbase3.apache.org,34991,1691649122119 in 170 msec 2023-08-10T06:32:39,128 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=37, resume processing ppid=36 2023-08-10T06:32:39,128 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=37, ppid=36, state=SUCCESS; TransitRegionStateProcedure table=table1, region=95cd117c0618eb0283c218ddc0657261, ASSIGN in 329 msec 2023-08-10T06:32:39,128 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=36, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:39,128 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649159128"}]},"ts":"1691649159128"} 2023-08-10T06:32:39,130 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=table1, state=ENABLED in hbase:meta 2023-08-10T06:32:39,133 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(81): pid=36, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:39,135 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=36, state=SUCCESS; CreateTableProcedure table=table1 in 1.1800 sec 2023-08-10T06:32:39,308 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_table1 2023-08-10T06:32:39,308 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_table1 Metrics about Tables on a single HBase RegionServer 2023-08-10T06:32:40,064 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=36 2023-08-10T06:32:40,064 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: CREATE, Table Name: default:table1, procId: 36 completed 2023-08-10T06:32:40,070 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.MapReduceRestoreJob(64): Restore full backup from directory hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3 from hbase tables test-1691649135102 to tables table1 2023-08-10T06:32:40,070 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.MapReduceRestoreJob(72): Restore test-1691649135102 into table1 2023-08-10T06:32:40,072 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.MapReduceHFileSplitterJob(107): add incremental job :hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/bulk_output-default-table1-1691649160070 from hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:40,075 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x65760690 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:40,081 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@f146ba1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:40,082 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:40,083 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:41924, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:32:40,097 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.HFileOutputFormat2(672): bulkload locality sensitive enabled 2023-08-10T06:32:40,097 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.HFileOutputFormat2(535): Looking up current regions for table table1 2023-08-10T06:32:40,102 DEBUG [hconnection-0x6d970dc4-shared-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:40,103 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:46614, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:40,106 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.HFileOutputFormat2(544): SplitPoint startkey for table1: 2023-08-10T06:32:40,107 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.HFileOutputFormat2(694): Configuring 1 reduce partitions to match current region count for all tables 2023-08-10T06:32:40,107 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.HFileOutputFormat2(559): Writing partition information to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/partitions_3cbfcf59-4d15-4069-a0be-d1410293f8c1 2023-08-10T06:32:40,373 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2023-08-10T06:32:40,376 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:52474, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2023-08-10T06:32:40,427 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'table1' 2023-08-10T06:32:40,546 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:40,546 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:40,547 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:40,547 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,592 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop-4209787289343533747.jar 2023-08-10T06:32:41,592 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,592 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,593 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,594 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,594 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,594 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,595 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,595 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,595 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,596 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.4/hbase-shaded-miscellaneous-4.1.4.jar 2023-08-10T06:32:41,597 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.4/hbase-shaded-gson-4.1.4.jar 2023-08-10T06:32:41,597 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.4/hbase-shaded-protobuf-4.1.4.jar 2023-08-10T06:32:41,598 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.4/hbase-shaded-netty-4.1.4.jar 2023-08-10T06:32:41,598 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.4/hbase-unsafe-4.1.4.jar 2023-08-10T06:32:41,599 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.5.7/zookeeper-3.5.7.jar 2023-08-10T06:32:41,599 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2023-08-10T06:32:41,599 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2023-08-10T06:32:41,600 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2023-08-10T06:32:41,600 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2023-08-10T06:32:41,601 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2023-08-10T06:32:41,601 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2023-08-10T06:32:41,605 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.io.ImmutableBytesWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,605 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.util.MapReduceExtendedCell, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,605 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.HFileInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,606 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.io.ImmutableBytesWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,606 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.util.MapReduceExtendedCell, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,607 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:41,607 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.10.2/hadoop-mapreduce-client-core-2.10.2.jar 2023-08-10T06:32:41,608 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.HFileOutputFormat2(714): Incremental output configured for tables: table1 2023-08-10T06:32:41,608 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:32:41,608 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x65760690 to 127.0.0.1:52413 2023-08-10T06:32:41,608 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:41,608 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.MapReduceHFileSplitterJob(119): success configuring load incremental job 2023-08-10T06:32:41,608 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(796): The addDependencyJars(Configuration, Class...) method has been deprecated since it is easy to use incorrectly. Most users should rely on addDependencyJars(Job) instead. See HBASE-8386 for more details. 2023-08-10T06:32:41,609 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.common.base.Preconditions, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.4/hbase-shaded-miscellaneous-4.1.4.jar 2023-08-10T06:32:41,615 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.JobResourceUploader(142): Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this. 2023-08-10T06:32:41,622 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.JobResourceUploader(470): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2023-08-10T06:32:41,711 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] fs.FileUtil(888): Command 'ln -s /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.local.dir/1691649161672/libjars /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/*' failed 1 with: ln: failed to create symbolic link '/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/*': No such file or directory 2023-08-10T06:32:41,711 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] mapred.LocalDistributedCacheManager(202): Failed to create symlink: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.local.dir/1691649161672/libjars <- /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/* 2023-08-10T06:32:41,783 INFO [LocalJobRunner Map Task Executor #0 {}] mapreduce.HFileInputFormat$HFileRecordReader(88): Initialize HFileRecordReader for hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:41,788 INFO [LocalJobRunner Map Task Executor #0 {}] mapreduce.HFileInputFormat$HFileRecordReader(101): Seeking to start 2023-08-10T06:32:41,793 DEBUG [LocalJobRunner Map Task Executor #0 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:41,925 DEBUG [pool-318-thread-1 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x2d3938ac to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:41,934 DEBUG [pool-318-thread-1 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4e422684, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:41,936 DEBUG [hconnection-0x4b361fed-metaLookup-shared--pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:41,938 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:35776, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:41,941 DEBUG [pool-318-thread-1 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x2d3938ac to 127.0.0.1:52413 2023-08-10T06:32:41,941 DEBUG [pool-318-thread-1 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:41,944 DEBUG [pool-318-thread-1 {}] mapreduce.HFileOutputFormat2$1(339): First rowkey: [row0] 2023-08-10T06:32:41,944 DEBUG [pool-318-thread-1 {}] mapreduce.HFileOutputFormat2$1(346): Use favored nodes writer: jenkins-hbase3.apache.org 2023-08-10T06:32:41,964 DEBUG [pool-318-thread-1 {}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/bulk_output-default-table1-1691649160070/_temporary/0/_temporary/attempt_local415538861_0002_r_000000_0/f/7dc66a0f98e848169f913848cfb90e45 is 37, key is row10/f:q1/1691649137773/Put/seqid=0 2023-08-10T06:32:41,970 WARN [Thread-988 {}] hdfs.DataStreamer(1828): These favored nodes were specified but not chosen: [jenkins-hbase3.apache.org:34991] Specified favored nodes: [jenkins-hbase3.apache.org:34991] 2023-08-10T06:32:42,766 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.MapReduceRestoreJob(89): Restoring HFiles from directory hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/bulk_output-default-table1-1691649160070 2023-08-10T06:32:42,766 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x0d680f65 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:42,775 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@626d3c52, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:42,777 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:42,779 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:35786, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:42,788 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] tool.LoadIncrementalHFiles(1067): Skipping non-directory hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/bulk_output-default-table1-1691649160070/_SUCCESS 2023-08-10T06:32:42,793 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:42,795 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:49296, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:32:42,826 DEBUG [LoadIncrementalHFiles-0 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 7dc66a0f98e848169f913848cfb90e45 2023-08-10T06:32:42,826 INFO [LoadIncrementalHFiles-0 {}] tool.LoadIncrementalHFiles(794): Trying to load hfile=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/bulk_output-default-table1-1691649160070/f/7dc66a0f98e848169f913848cfb90e45 first=Optional[row0] last=Optional[row98] 2023-08-10T06:32:42,851 DEBUG [LoadIncrementalHFiles-0 {}] tool.LoadIncrementalHFiles$1(559): Going to connect to server region=table1,,1691649157953.95cd117c0618eb0283c218ddc0657261., hostname=jenkins-hbase3.apache.org,34991,1691649122119, seqNum=2 for row with hfile group [{f,hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/bulk_output-default-table1-1691649160070/f/7dc66a0f98e848169f913848cfb90e45}] 2023-08-10T06:32:42,872 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(611): Validating hfile at hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/bulk_output-default-table1-1691649160070/f/7dc66a0f98e848169f913848cfb90e45 for inclusion in 95cd117c0618eb0283c218ddc0657261/f 2023-08-10T06:32:42,881 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(623): HFile bounds: first=row0 last=row98 2023-08-10T06:32:42,881 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(625): Region bounds: first= last= 2023-08-10T06:32:42,881 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 7dc66a0f98e848169f913848cfb90e45 2023-08-10T06:32:42,883 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HRegion(2520): Flush status journal for 95cd117c0618eb0283c218ddc0657261: 2023-08-10T06:32:42,884 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.SecureBulkLoadManager$SecureBulkLoadListener(397): Moving hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/hbase-staging/bulk_output-default-table1-1691649160070/f/7dc66a0f98e848169f913848cfb90e45 to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__292vpj5etl37j4vjsslrlhs3r02ed3vsml62i6mhh6nthatdoh23o8o7phdduihd/f/7dc66a0f98e848169f913848cfb90e45 2023-08-10T06:32:42,889 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__292vpj5etl37j4vjsslrlhs3r02ed3vsml62i6mhh6nthatdoh23o8o7phdduihd/f/7dc66a0f98e848169f913848cfb90e45 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/f/7cacdec94e2f4d84a2b23b96215c0cda_SeqId_4_ 2023-08-10T06:32:42,890 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x3c4d428b to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:42,895 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7e0dc4bf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:42,896 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:42,898 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:49310, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:32:42,902 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:42,904 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:35798, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:42,924 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:32:42,924 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x3c4d428b to 127.0.0.1:52413 2023-08-10T06:32:42,924 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:42,926 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(697): Loaded HFile hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__292vpj5etl37j4vjsslrlhs3r02ed3vsml62i6mhh6nthatdoh23o8o7phdduihd/f/7dc66a0f98e848169f913848cfb90e45 into 95cd117c0618eb0283c218ddc0657261/f as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/f/7cacdec94e2f4d84a2b23b96215c0cda_SeqId_4_ - updating store file list. 2023-08-10T06:32:42,932 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(719): Loaded HFile hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/f/7cacdec94e2f4d84a2b23b96215c0cda_SeqId_4_ into 95cd117c0618eb0283c218ddc0657261/f 2023-08-10T06:32:42,932 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(703): Successfully loaded hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__292vpj5etl37j4vjsslrlhs3r02ed3vsml62i6mhh6nthatdoh23o8o7phdduihd/f/7dc66a0f98e848169f913848cfb90e45 into 95cd117c0618eb0283c218ddc0657261/f (new location: hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/f/7cacdec94e2f4d84a2b23b96215c0cda_SeqId_4_) 2023-08-10T06:32:42,933 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.SecureBulkLoadManager$SecureBulkLoadListener(412): Bulk Load done for: hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__292vpj5etl37j4vjsslrlhs3r02ed3vsml62i6mhh6nthatdoh23o8o7phdduihd/f/7dc66a0f98e848169f913848cfb90e45 2023-08-10T06:32:42,940 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x44e25f6b to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:42,947 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@439200ee, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:42,948 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:42,949 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:49318, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=MasterService 2023-08-10T06:32:42,952 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:42,953 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:35800, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=ClientService 2023-08-10T06:32:42,967 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:32:42,968 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x44e25f6b to 127.0.0.1:52413 2023-08-10T06:32:42,968 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:42,977 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:32:42,977 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x0d680f65 to 127.0.0.1:52413 2023-08-10T06:32:42,977 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:42,978 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.MapReduceRestoreJob(100): Restore Job finished:0 2023-08-10T06:32:42,978 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.RestoreTablesClient(231): Restore includes the following image(s): 2023-08-10T06:32:42,978 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.RestoreTablesClient(233): Backup: backup_1691649143378 hdfs://localhost:44455/backupUT/backup_1691649143378/default/test-1691649135102/ 2023-08-10T06:32:42,978 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.RestoreTablesClient(238): restoreStage finished 2023-08-10T06:32:42,981 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$18(963): Started disable of table1 2023-08-10T06:32:42,986 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$12(2663): Client=jenkins//172.31.12.81 disable table1 2023-08-10T06:32:42,995 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=39, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=table1 2023-08-10T06:32:42,999 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649162998"}]},"ts":"1691649162998"} 2023-08-10T06:32:43,000 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=39 2023-08-10T06:32:43,000 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=table1, state=DISABLING in hbase:meta 2023-08-10T06:32:43,002 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(293): Set table1 to state=DISABLING 2023-08-10T06:32:43,003 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=40, ppid=39, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=table1, region=95cd117c0618eb0283c218ddc0657261, UNASSIGN}] 2023-08-10T06:32:43,005 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=40, ppid=39, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=table1, region=95cd117c0618eb0283c218ddc0657261, UNASSIGN 2023-08-10T06:32:43,006 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=40 updating hbase:meta row=95cd117c0618eb0283c218ddc0657261, regionState=CLOSING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:43,007 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=41, ppid=40, state=RUNNABLE; CloseRegionProcedure 95cd117c0618eb0283c218ddc0657261, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:43,101 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=39 2023-08-10T06:32:43,167 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] handler.UnassignRegionHandler(119): Close 95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:43,167 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] regionserver.HRegion(1668): Closing 95cd117c0618eb0283c218ddc0657261, disabling compactions & flushes 2023-08-10T06:32:43,167 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] regionserver.HRegion(1690): Closing region table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:43,167 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] regionserver.HRegion(1711): Waiting without time limit for close lock on table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:43,167 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] regionserver.HRegion(1778): Acquired close lock on table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. after waiting 0 ms 2023-08-10T06:32:43,167 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] regionserver.HRegion(1788): Updates disabled for region table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:43,170 DEBUG [StoreFileCloser-95cd117c0618eb0283c218ddc0657261-f-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 7cacdec94e2f4d84a2b23b96215c0cda_SeqId_4_ 2023-08-10T06:32:43,175 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/recovered.edits/6.seqid, newMaxSeqId=6, maxSeqId=1 2023-08-10T06:32:43,176 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:32:43,176 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] regionserver.HRegion(1909): Closed table1,,1691649157953.95cd117c0618eb0283c218ddc0657261. 2023-08-10T06:32:43,176 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] regionserver.HRegion(1622): Region close journal for 95cd117c0618eb0283c218ddc0657261: 2023-08-10T06:32:43,178 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=41}] handler.UnassignRegionHandler(163): Closed 95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:43,179 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=40 updating hbase:meta row=95cd117c0618eb0283c218ddc0657261, regionState=CLOSED 2023-08-10T06:32:43,182 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=41, resume processing ppid=40 2023-08-10T06:32:43,182 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=41, ppid=40, state=SUCCESS; CloseRegionProcedure 95cd117c0618eb0283c218ddc0657261, server=jenkins-hbase3.apache.org,34991,1691649122119 in 173 msec 2023-08-10T06:32:43,184 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=40, resume processing ppid=39 2023-08-10T06:32:43,184 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=40, ppid=39, state=SUCCESS; TransitRegionStateProcedure table=table1, region=95cd117c0618eb0283c218ddc0657261, UNASSIGN in 179 msec 2023-08-10T06:32:43,185 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649163184"}]},"ts":"1691649163184"} 2023-08-10T06:32:43,186 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=table1, state=DISABLED in hbase:meta 2023-08-10T06:32:43,188 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(305): Set table1 to state=DISABLED 2023-08-10T06:32:43,190 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=39, state=SUCCESS; DisableTableProcedure table=table1 in 201 msec 2023-08-10T06:32:43,302 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=39 2023-08-10T06:32:43,303 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: DISABLE, Table Name: default:table1, procId: 39 completed 2023-08-10T06:32:43,308 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$5(2447): Client=jenkins//172.31.12.81 delete table1 2023-08-10T06:32:43,316 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=42, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=table1 2023-08-10T06:32:43,318 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(98): Waiting for RIT for pid=42, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:32:43,319 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(110): Deleting regions from filesystem for pid=42, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:32:43,321 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=42 2023-08-10T06:32:43,325 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:43,330 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/f, FileablePath, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/recovered.edits] 2023-08-10T06:32:43,336 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/f/7cacdec94e2f4d84a2b23b96215c0cda_SeqId_4_ to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/archive/data/default/table1/95cd117c0618eb0283c218ddc0657261/f/7cacdec94e2f4d84a2b23b96215c0cda_SeqId_4_ 2023-08-10T06:32:43,340 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261/recovered.edits/6.seqid to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/archive/data/default/table1/95cd117c0618eb0283c218ddc0657261/recovered.edits/6.seqid 2023-08-10T06:32:43,341 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95cd117c0618eb0283c218ddc0657261 2023-08-10T06:32:43,341 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(293): Archived table1 regions 2023-08-10T06:32:43,343 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from META for pid=42, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:32:43,352 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(350): Deleting some vestigial 1 rows of table1 from hbase:meta 2023-08-10T06:32:43,354 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(387): Removing 'table1' descriptor. 2023-08-10T06:32:43,355 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(121): Deleting assignment state for pid=42, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:32:43,355 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(377): Removing 'table1' from region states. 2023-08-10T06:32:43,356 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"table1,,1691649157953.95cd117c0618eb0283c218ddc0657261.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1691649163355"}]},"ts":"9223372036854775807"} 2023-08-10T06:32:43,358 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 1 regions from META 2023-08-10T06:32:43,358 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 95cd117c0618eb0283c218ddc0657261, NAME => 'table1,,1691649157953.95cd117c0618eb0283c218ddc0657261.', STARTKEY => '', ENDKEY => ''}] 2023-08-10T06:32:43,358 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(381): Marking 'table1' as deleted. 2023-08-10T06:32:43,358 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1691649163358"}]},"ts":"9223372036854775807"} 2023-08-10T06:32:43,360 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table table1 state from META 2023-08-10T06:32:43,362 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(127): Finished pid=42, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:32:43,363 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=42, state=SUCCESS; DeleteTableProcedure table=table1 in 53 msec 2023-08-10T06:32:43,422 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=42 2023-08-10T06:32:43,422 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: DELETE, Table Name: default:table1, procId: 42 completed 2023-08-10T06:32:43,447 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.ResourceChecker(175): after: backup.TestRemoteRestore#testFullRestoreRemote Thread=831 (was 815) Potentially hanging thread: IPC Client (1886842325) connection to localhost/127.0.0.1:44455 from jenkins.hfs.1 java.lang.Object.wait(Native Method) org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1053) org.apache.hadoop.ipc.Client$Connection.run(Client.java:1097) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: PacketResponder: BP-2138422956-172.31.12.81-1691649115962:blk_1073741858_1034, type=LAST_IN_PIPELINE java.lang.Object.wait(Native Method) java.lang.Object.wait(Object.java:502) org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.waitForAckHead(BlockReceiver.java:1327) org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1399) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Close-WAL-Writer-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460) java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362) java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-7 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-11 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: RS-EventLoopGroup-5-3 org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:209) org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:202) org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:306) org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:363) org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: HFileArchiver-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-6 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/cluster_31ee86f8-f5ed-8a45-aa31-8b420ad4a168/dfs/data/data1/current sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_2129524288_16 at /127.0.0.1:42112 [Receiving block BP-2138422956-172.31.12.81-1691649115962:blk_1073741857_1033] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read1(BufferedInputStream.java:286) java.io.BufferedInputStream.read(BufferedInputStream.java:345) java.io.DataInputStream.read(DataInputStream.java:149) org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:209) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:211) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109) org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:528) org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:968) org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:877) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:166) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:103) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-8 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/cluster_31ee86f8-f5ed-8a45-aa31-8b420ad4a168/dfs/data/data2/current sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: PacketResponder: BP-2138422956-172.31.12.81-1691649115962:blk_1073741857_1033, type=LAST_IN_PIPELINE java.lang.Object.wait(Native Method) java.lang.Object.wait(Object.java:502) org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.waitForAckHead(BlockReceiver.java:1327) org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1399) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-10 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: (jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:458) java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362) java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:924) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_242875349_16 at /127.0.0.1:36554 [Waiting for operation #8] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read(BufferedInputStream.java:265) java.io.DataInputStream.readShort(DataInputStream.java:312) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:67) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:269) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1973487143_16 at /127.0.0.1:43658 [Waiting for operation #4] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read(BufferedInputStream.java:265) java.io.DataInputStream.readShort(DataInputStream.java:312) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:67) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:269) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_2129524288_16 at /127.0.0.1:42128 [Receiving block BP-2138422956-172.31.12.81-1691649115962:blk_1073741858_1034] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read1(BufferedInputStream.java:286) java.io.BufferedInputStream.read(BufferedInputStream.java:345) java.io.DataInputStream.read(DataInputStream.java:149) org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:209) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:211) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109) org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:528) org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:968) org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:877) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:166) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:103) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:458) java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362) java.util.concurrent.SynchronousQueue.take(SynchronousQueue.java:924) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-9 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Close-WAL-Writer-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460) java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362) java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) - Thread LEAK? -, OpenFileDescriptor=903 (was 878) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=60000 (was 60000), SystemLoadAverage=91 (was 93), ProcessCount=171 (was 169) - ProcessCount LEAK? -, AvailableMemoryMB=1985 (was 2357) 2023-08-10T06:32:43,448 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.ResourceChecker(130): Thread=831 is superior to 500 2023-08-10T06:32:43,467 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.ResourceChecker(147): before: backup.TestRemoteRestore#testFullRestoreRemoteWithAlternateRestoreOutputDir Thread=831, OpenFileDescriptor=903, MaxFileDescriptor=60000, SystemLoadAverage=91, ProcessCount=171, AvailableMemoryMB=1984 2023-08-10T06:32:43,467 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.ResourceChecker(130): Thread=831 is superior to 500 2023-08-10T06:32:43,467 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] backup.TestRemoteRestore(91): test remote full backup on a single table with alternate restore output dir 2023-08-10T06:32:43,468 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x0d3a42bc to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:43,473 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6981c151, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:43,476 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:43,478 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:35816, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:43,480 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:43,481 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:49320, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:32:43,494 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(610): Start new backup exclusive operation 2023-08-10T06:32:43,500 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1685): Client=jenkins//172.31.12.81 snapshot request for:{ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } 2023-08-10T06:32:43,500 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1691649163500 (current time:1691649163500). 2023-08-10T06:32:43,500 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2023-08-10T06:32:43,500 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snapshot_backup_system VERSION not specified, setting to 2 2023-08-10T06:32:43,500 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x7fbe1343 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:43,505 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2cf82e0c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:43,506 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:43,508 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:35820, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:43,509 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x7fbe1343 to 127.0.0.1:52413 2023-08-10T06:32:43,509 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:43,509 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(770): No existing snapshot, attempting snapshot... 2023-08-10T06:32:43,510 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=43, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } 2023-08-10T06:32:43,510 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(1404): register snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 }, snapshot procedure id = 43 2023-08-10T06:32:43,512 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2023-08-10T06:32:43,512 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=43 2023-08-10T06:32:43,512 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2023-08-10T06:32:43,514 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2023-08-10T06:32:43,614 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=43 2023-08-10T06:32:43,815 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=43 2023-08-10T06:32:43,923 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2023-08-10T06:32:43,924 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=44, ppid=43, state=RUNNABLE; SnapshotRegionProcedure de2170c3441572f6a0fca17a49c93cc1}] 2023-08-10T06:32:43,926 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=44, ppid=43, state=RUNNABLE; SnapshotRegionProcedure de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:44,077 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=34991 {}] regionserver.RSRpcServices(3924): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=44 2023-08-10T06:32:44,078 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:44,078 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.HRegion(2819): Flushing de2170c3441572f6a0fca17a49c93cc1 2/2 column families, dataSize=1.08 KB heapSize=2.23 KB 2023-08-10T06:32:44,092 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/meta/53df5bd3efe64cb28d0a568f4284b10b is 169, key is trslm:hdfs://localhost:44455/backupUT\x00test-1691649135102/meta:log-roll-map/1691649157436/Put/seqid=0 2023-08-10T06:32:44,116 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=43 2023-08-10T06:32:44,497 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=526 B at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/meta/53df5bd3efe64cb28d0a568f4284b10b 2023-08-10T06:32:44,504 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 53df5bd3efe64cb28d0a568f4284b10b 2023-08-10T06:32:44,507 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/397d772dc94c4e45be47d9f2f04a3258 is 310, key is session:backup_1691649143378/session:context/1691649157896/Put/seqid=0 2023-08-10T06:32:44,617 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=43 2023-08-10T06:32:44,716 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2023-08-10T06:32:44,915 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=579 B at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/397d772dc94c4e45be47d9f2f04a3258 2023-08-10T06:32:44,923 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 397d772dc94c4e45be47d9f2f04a3258 2023-08-10T06:32:44,925 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/meta/53df5bd3efe64cb28d0a568f4284b10b as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/meta/53df5bd3efe64cb28d0a568f4284b10b 2023-08-10T06:32:44,933 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/meta/53df5bd3efe64cb28d0a568f4284b10b, entries=4, sequenceid=17, filesize=5.5 K 2023-08-10T06:32:44,934 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/397d772dc94c4e45be47d9f2f04a3258 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/397d772dc94c4e45be47d9f2f04a3258 2023-08-10T06:32:44,941 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/397d772dc94c4e45be47d9f2f04a3258, entries=2, sequenceid=17, filesize=5.3 K 2023-08-10T06:32:44,942 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.HRegion(3022): Finished flush of dataSize ~1.08 KB/1105, heapSize ~2.20 KB/2256, currentSize=0 B/0 for de2170c3441572f6a0fca17a49c93cc1 in 864ms, sequenceid=17, compaction requested=false 2023-08-10T06:32:44,942 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.HRegion(2520): Flush status journal for de2170c3441572f6a0fca17a49c93cc1: 2023-08-10T06:32:44,942 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.SnapshotRegionCallable(75): Snapshotting region backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. for snapshot_backup_system completed. 2023-08-10T06:32:44,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] snapshot.SnapshotManifest(240): Storing 'backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1.' region-info for snapshot=snapshot_backup_system 2023-08-10T06:32:44,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] snapshot.SnapshotManifest(245): Creating references for hfiles 2023-08-10T06:32:44,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] snapshot.SnapshotManifest(255): Adding snapshot references for [hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/meta/53df5bd3efe64cb28d0a568f4284b10b] hfiles 2023-08-10T06:32:44,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] snapshot.SnapshotManifest(264): Adding reference for file (1/1): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/meta/53df5bd3efe64cb28d0a568f4284b10b for snapshot=snapshot_backup_system 2023-08-10T06:32:44,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] snapshot.SnapshotManifest(255): Adding snapshot references for [hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/a7aa6b78efc846bebb42d595b96dfe44, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/397d772dc94c4e45be47d9f2f04a3258] hfiles 2023-08-10T06:32:44,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] snapshot.SnapshotManifest(264): Adding reference for file (1/2): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/a7aa6b78efc846bebb42d595b96dfe44 for snapshot=snapshot_backup_system 2023-08-10T06:32:44,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] snapshot.SnapshotManifest(264): Adding reference for file (2/2): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/397d772dc94c4e45be47d9f2f04a3258 for snapshot=snapshot_backup_system 2023-08-10T06:32:45,353 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:32:45,353 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=44}] regionserver.RemoteProcedureResultReporter(61): Successfully complete execution of pid=44 2023-08-10T06:32:45,353 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster(3994): Remote procedure done, pid=44 2023-08-10T06:32:45,353 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure.SnapshotRegionProcedure(130): finish snapshot snapshot_backup_system on region de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:45,355 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=44, ppid=43, state=RUNNABLE; SnapshotRegionProcedure de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:32:45,358 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=44, resume processing ppid=43 2023-08-10T06:32:45,358 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2023-08-10T06:32:45,358 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=44, ppid=43, state=SUCCESS; SnapshotRegionProcedure de2170c3441572f6a0fca17a49c93cc1 in 1.4330 sec 2023-08-10T06:32:45,359 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2023-08-10T06:32:45,359 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2023-08-10T06:32:45,359 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(489): Convert to Single Snapshot Manifest for snapshot_backup_system 2023-08-10T06:32:45,360 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(126): No regions under directory:hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/.tmp/snapshot_backup_system 2023-08-10T06:32:45,618 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=43 2023-08-10T06:32:45,771 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2023-08-10T06:32:45,778 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2023-08-10T06:32:45,778 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/.tmp/snapshot_backup_system to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/snapshot_backup_system 2023-08-10T06:32:46,597 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=43, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2023-08-10T06:32:46,597 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1410): unregister snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 }, snapshot procedure id = 43 2023-08-10T06:32:46,599 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=43, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=43, snapshot={ ss=snapshot_backup_system table=backup:system type=FLUSH ttl=0 } in 3.0880 sec 2023-08-10T06:32:47,590 INFO [master/jenkins-hbase3:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2023-08-10T06:32:47,590 INFO [master/jenkins-hbase3:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2023-08-10T06:32:47,619 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=43 2023-08-10T06:32:47,620 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: SNAPSHOT, Table Name: backup:system, procId: 43 completed 2023-08-10T06:32:47,620 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(115): Backup backup_1691649163473 started at 1691649167620. 2023-08-10T06:32:47,622 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(119): Backup session backup_1691649163473 has been started. 2023-08-10T06:32:47,627 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(150): Execute roll log procedure for full backup ... 2023-08-10T06:32:47,628 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(946): Client=jenkins//172.31.12.81 procedure request for: rolllog-proc 2023-08-10T06:32:47,628 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure.ProcedureCoordinator(165): Submitting procedure rolllog 2023-08-10T06:32:47,629 INFO [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(191): Starting procedure 'rolllog' 2023-08-10T06:32:47,629 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] errorhandling.TimeoutExceptionInjector(107): Scheduling process timer to run in: 180000 ms 2023-08-10T06:32:47,630 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(199): Procedure 'rolllog' starting 'acquire' 2023-08-10T06:32:47,630 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(241): Starting procedure 'rolllog', kicking off acquire phase on members. 2023-08-10T06:32:47,631 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,631 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureCoordinator(92): Creating acquire znode:/1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:47,632 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/acquired 2023-08-10T06:32:47,632 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureCoordinator(100): Watching for acquire node:/1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,632 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(104): Received procedure start children changed event: /1/rolllog-proc/acquired 2023-08-10T06:32:47,632 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/1/rolllog-proc/acquired' 2023-08-10T06:32:47,632 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,632 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(203): Waiting for all members to 'acquire' 2023-08-10T06:32:47,632 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(186): Found procedure znode: /1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:47,633 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,633 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(212): start proc data length is 35 2023-08-10T06:32:47,633 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(214): Found data for znode:/1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:47,633 INFO [zk-event-processor-pool-0 {}] regionserver.LogRollRegionServerProcedureManager(126): Attempting to run a roll log procedure for backup. 2023-08-10T06:32:47,633 INFO [zk-event-processor-pool-0 {}] regionserver.LogRollBackupSubprocedure(56): Constructing a LogRollBackupSubprocedure. 2023-08-10T06:32:47,633 DEBUG [zk-event-processor-pool-0 {}] procedure.ProcedureMember(140): Submitting new Subprocedure:rolllog 2023-08-10T06:32:47,633 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(151): Starting subprocedure 'rolllog' with timeout 60000ms 2023-08-10T06:32:47,633 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] errorhandling.TimeoutExceptionInjector(107): Scheduling process timer to run in: 60000 ms 2023-08-10T06:32:47,634 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(159): Subprocedure 'rolllog' starting 'acquire' stage 2023-08-10T06:32:47,634 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(161): Subprocedure 'rolllog' locally acquired 2023-08-10T06:32:47,635 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.ZKProcedureMemberRpcs(242): Member: 'jenkins-hbase3.apache.org,34991,1691649122119' joining acquired barrier for procedure (rolllog) in zk 2023-08-10T06:32:47,636 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.ZKProcedureMemberRpcs(250): Watch for global barrier reached:/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:47,636 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,636 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureCoordinator$1(194): Node created: /1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,636 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(225): Current zk system: 2023-08-10T06:32:47,636 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(227): |-/1/rolllog-proc 2023-08-10T06:32:47,636 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/reached/rolllog 2023-08-10T06:32:47,636 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(166): Subprocedure 'rolllog' coordinator notified of 'acquire', waiting on 'reached' or 'abort' from coordinator 2023-08-10T06:32:47,636 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-abort 2023-08-10T06:32:47,637 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-acquired 2023-08-10T06:32:47,637 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:47,637 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,637 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-reached 2023-08-10T06:32:47,638 DEBUG [zk-event-processor-pool-0 {}] procedure.Procedure(291): member: 'jenkins-hbase3.apache.org,34991,1691649122119' joining acquired barrier for procedure 'rolllog' on coordinator 2023-08-10T06:32:47,638 DEBUG [zk-event-processor-pool-0 {}] procedure.Procedure(300): Waiting on: java.util.concurrent.CountDownLatch@1018d1f2[Count = 0] remaining members to acquire global barrier 2023-08-10T06:32:47,638 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(207): Procedure 'rolllog' starting 'in-barrier' execution. 2023-08-10T06:32:47,638 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureCoordinator(116): Creating reached barrier zk node:/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:47,639 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:47,639 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(77): Received created event:/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:47,639 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(128): Received reached global barrier:/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:47,639 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(180): Subprocedure 'rolllog' received 'reached' from coordinator. 2023-08-10T06:32:47,639 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,639 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(211): Waiting for all members to 'release' 2023-08-10T06:32:47,639 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] regionserver.LogRollBackupSubprocedurePool(84): Waiting for backup procedure to finish. 2023-08-10T06:32:47,639 DEBUG [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] regionserver.LogRollBackupSubprocedure$RSRollLogTask(75): DRPC started: jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,639 INFO [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] regionserver.LogRollBackupSubprocedure$RSRollLogTask(92): Trying to roll log in backup subprocedure, current log number: 1691649150558 highest: 1691649150558 on jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,640 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractWALRoller(197): WAL AsyncFSWAL jenkins-hbase3.apache.org%2C34991%2C1691649122119:(num 1691649150558) roll requested 2023-08-10T06:32:47,652 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649167640, exclude list is [], retry=0 2023-08-10T06:32:47,657 DEBUG [RS-EventLoopGroup-3-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK] 2023-08-10T06:32:47,659 INFO [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(828): Rolled WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649150558 with entries=15, filesize=3.47 KB; new WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649167640 2023-08-10T06:32:47,659 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK]] 2023-08-10T06:32:47,659 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(742): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.1691649150558 is not closed yet, will try archiving it next time 2023-08-10T06:32:47,659 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractWALRoller(197): WAL AsyncFSWAL jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta:.meta(num 1691649150584) roll requested 2023-08-10T06:32:47,673 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(531): When create output stream for /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649167659.meta, exclude list is [], retry=0 2023-08-10T06:32:47,680 DEBUG [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] regionserver.LogRollBackupSubprocedure$RSRollLogTask(99): log roll took 41 2023-08-10T06:32:47,680 DEBUG [RS-EventLoopGroup-3-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(810): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK] 2023-08-10T06:32:47,680 INFO [rs(jenkins-hbase3.apache.org,34991,1691649122119)-backup-pool-0 {}] regionserver.LogRollBackupSubprocedure$RSRollLogTask(100): After roll log in backup subprocedure, current log number: 1691649167640 on jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,681 INFO [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(828): Rolled WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649150584.meta with entries=12, filesize=2.73 KB; new WAL /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649167659.meta 2023-08-10T06:32:47,682 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(916): Create new AsyncFSWAL writer with pipeline: [DatanodeInfoWithStorage[127.0.0.1:45413,DS-2b629751-7c8a-4eca-b84d-f781cc15a0d0,DISK]] 2023-08-10T06:32:47,682 DEBUG [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractFSWAL(742): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/WALs/jenkins-hbase3.apache.org,34991,1691649122119/jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta.1691649150584.meta is not closed yet, will try archiving it next time 2023-08-10T06:32:47,692 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(182): Subprocedure 'rolllog' locally completed 2023-08-10T06:32:47,692 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.ZKProcedureMemberRpcs(267): Marking procedure 'rolllog' completed for member 'jenkins-hbase3.apache.org,34991,1691649122119' in zk 2023-08-10T06:32:47,694 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(187): Subprocedure 'rolllog' has notified controller of completion 2023-08-10T06:32:47,694 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,694 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] errorhandling.TimeoutExceptionInjector(87): Marking timer as complete - no error notifications will be received for this timer. 2023-08-10T06:32:47,695 DEBUG [member: 'jenkins-hbase3.apache.org,34991,1691649122119' subprocedure-pool-0 {}] procedure.Subprocedure(212): Subprocedure 'rolllog' completed. 2023-08-10T06:32:47,696 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureCoordinator$1(194): Node created: /1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,696 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(225): Current zk system: 2023-08-10T06:32:47,696 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(227): |-/1/rolllog-proc 2023-08-10T06:32:47,696 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-abort 2023-08-10T06:32:47,696 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-acquired 2023-08-10T06:32:47,697 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:47,697 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,697 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-reached 2023-08-10T06:32:47,698 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:47,698 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,698 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureCoordinator$1(218): Finished data from procedure 'rolllog' member 'jenkins-hbase3.apache.org,34991,1691649122119': 2023-08-10T06:32:47,699 DEBUG [zk-event-processor-pool-0 {}] procedure.Procedure(321): Member: 'jenkins-hbase3.apache.org,34991,1691649122119' released barrier for procedure'rolllog', counting down latch. Waiting for 0 more 2023-08-10T06:32:47,699 INFO [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(216): Procedure 'rolllog' execution completed 2023-08-10T06:32:47,699 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(225): Running finish phase. 2023-08-10T06:32:47,699 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.Procedure(275): Finished coordinator procedure - removing self from list of running procedures 2023-08-10T06:32:47,699 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureCoordinator(162): Attempting to clean out zk node for op:rolllog 2023-08-10T06:32:47,699 INFO [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] procedure.ZKProcedureUtil(265): Clearing all znodes for procedure rolllogincluding nodes /1/rolllog-proc/acquired /1/rolllog-proc/reached /1/rolllog-proc/abort 2023-08-10T06:32:47,700 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,700 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,700 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/abort 2023-08-10T06:32:47,700 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(77): Received created event:/1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,700 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureCoordinator$1(194): Node created: /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,700 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(316): Aborting procedure member for znode /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,700 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(225): Current zk system: 2023-08-10T06:32:47,700 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(227): |-/1/rolllog-proc 2023-08-10T06:32:47,701 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(107): Received procedure abort children changed event: /1/rolllog-proc/abort 2023-08-10T06:32:47,701 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-abort 2023-08-10T06:32:47,701 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/1/rolllog-proc/abort' 2023-08-10T06:32:47,701 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(111): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on existing znode=/1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,701 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(316): Aborting procedure member for znode /1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,701 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:47,702 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-acquired 2023-08-10T06:32:47,702 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:47,702 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,703 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] zookeeper.ZKUtil(111): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on existing znode=/1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,703 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-reached 2023-08-10T06:32:47,703 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |----rolllog 2023-08-10T06:32:47,703 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureUtil(244): |-------jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,705 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/acquired 2023-08-10T06:32:47,705 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/acquired/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,705 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:47,705 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(104): Received procedure start children changed event: /1/rolllog-proc/acquired 2023-08-10T06:32:47,706 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/1/rolllog-proc/acquired' 2023-08-10T06:32:47,706 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/acquired/rolllog 2023-08-10T06:32:47,706 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:47,706 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/abort 2023-08-10T06:32:47,706 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:47,706 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/reached/rolllog 2023-08-10T06:32:47,706 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rolllog-proc/abort/rolllog 2023-08-10T06:32:47,706 DEBUG [(jenkins-hbase3.apache.org,42279,1691649119143)-proc-coordinator-pool-0 {}] errorhandling.TimeoutExceptionInjector(87): Marking timer as complete - no error notifications will be received for this timer. 2023-08-10T06:32:47,706 INFO [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs$1(107): Received procedure abort children changed event: /1/rolllog-proc/abort 2023-08-10T06:32:47,706 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.LogRollMasterProcedureManager(143): Done waiting - exec procedure for rolllog 2023-08-10T06:32:47,707 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.LogRollMasterProcedureManager(144): Distributed roll log procedure is successful! 2023-08-10T06:32:47,707 DEBUG [zk-event-processor-pool-0 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/1/rolllog-proc/abort' 2023-08-10T06:32:47,707 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin(2751): Waiting a max of 300000 ms for procedure 'rolllog-proc : rolllog'' to complete. (max 6666 ms per retry) 2023-08-10T06:32:47,707 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin(2759): (#1) Sleeping: 100ms while waiting for procedure completion. 2023-08-10T06:32:47,808 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin(2765): Getting current status of procedure from master... 2023-08-10T06:32:47,809 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1247): Checking to see if procedure from request:rolllog-proc is done 2023-08-10T06:32:47,812 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1685): Client=jenkins//172.31.12.81 snapshot request for:{ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } 2023-08-10T06:32:47,812 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1691649167812 (current time:1691649167812). 2023-08-10T06:32:47,812 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2023-08-10T06:32:47,812 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snapshot_1691649167811_default_test-1691649135102 VERSION not specified, setting to 2 2023-08-10T06:32:47,813 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x59d92292 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:47,823 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6c08cf5a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:47,824 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:47,826 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:35828, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:47,827 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x59d92292 to 127.0.0.1:52413 2023-08-10T06:32:47,827 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:47,828 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(770): No existing snapshot, attempting snapshot... 2023-08-10T06:32:47,829 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=45, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } 2023-08-10T06:32:47,829 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(1404): register snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 }, snapshot procedure id = 45 2023-08-10T06:32:47,831 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2023-08-10T06:32:47,831 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=45 2023-08-10T06:32:47,832 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2023-08-10T06:32:47,833 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2023-08-10T06:32:47,933 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=45 2023-08-10T06:32:48,134 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=45 2023-08-10T06:32:48,240 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2023-08-10T06:32:48,240 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=46, ppid=45, state=RUNNABLE; SnapshotRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3}] 2023-08-10T06:32:48,242 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=46, ppid=45, state=RUNNABLE; SnapshotRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:48,393 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=34991 {}] regionserver.RSRpcServices(3924): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=46 2023-08-10T06:32:48,394 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:48,394 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] regionserver.HRegion(2520): Flush status journal for 7f5405ff46e3916fa2d5dcf037138bc3: 2023-08-10T06:32:48,394 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] regionserver.SnapshotRegionCallable(75): Snapshotting region test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. for snapshot_1691649167811_default_test-1691649135102 completed. 2023-08-10T06:32:48,394 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] snapshot.SnapshotManifest(240): Storing 'test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3.' region-info for snapshot=snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:48,394 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] snapshot.SnapshotManifest(245): Creating references for hfiles 2023-08-10T06:32:48,394 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] snapshot.SnapshotManifest(255): Adding snapshot references for [hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de] hfiles 2023-08-10T06:32:48,394 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] snapshot.SnapshotManifest(264): Adding reference for file (1/1): hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de for snapshot=snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:48,435 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=45 2023-08-10T06:32:48,800 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:32:48,801 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=46}] regionserver.RemoteProcedureResultReporter(61): Successfully complete execution of pid=46 2023-08-10T06:32:48,801 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster(3994): Remote procedure done, pid=46 2023-08-10T06:32:48,801 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure.SnapshotRegionProcedure(130): finish snapshot snapshot_1691649167811_default_test-1691649135102 on region 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:48,803 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=46, ppid=45, state=RUNNABLE; SnapshotRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:48,805 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=46, resume processing ppid=45 2023-08-10T06:32:48,805 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2023-08-10T06:32:48,805 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=46, ppid=45, state=SUCCESS; SnapshotRegionProcedure 7f5405ff46e3916fa2d5dcf037138bc3 in 564 msec 2023-08-10T06:32:48,806 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2023-08-10T06:32:48,806 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2023-08-10T06:32:48,806 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(489): Convert to Single Snapshot Manifest for snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:48,807 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(126): No regions under directory:hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/.tmp/snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:48,936 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=45 2023-08-10T06:32:49,217 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2023-08-10T06:32:49,224 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2023-08-10T06:32:49,224 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/.tmp/snapshot_1691649167811_default_test-1691649135102 to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:49,308 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_table1 2023-08-10T06:32:49,643 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=45, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2023-08-10T06:32:49,643 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1410): unregister snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 }, snapshot procedure id = 45 2023-08-10T06:32:49,644 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=45, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=45, snapshot={ ss=snapshot_1691649167811_default_test-1691649135102 table=test-1691649135102 type=FLUSH ttl=0 } in 1.8150 sec 2023-08-10T06:32:49,937 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=45 2023-08-10T06:32:49,938 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.HBaseAdmin$TableFuture(3590): Operation: SNAPSHOT, Table Name: default:test-1691649135102, procId: 45 completed 2023-08-10T06:32:49,938 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(171): snapshot copy for backup_1691649163473 2023-08-10T06:32:49,938 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(68): Snapshot copy is starting. 2023-08-10T06:32:49,938 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(80): There are 1 snapshots to be copied. 2023-08-10T06:32:49,938 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(105): Setting snapshot copy job name to : Full-Backup_backup_1691649163473_test-1691649135102 2023-08-10T06:32:49,938 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(109): Copy snapshot snapshot_1691649167811_default_test-1691649135102 to hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/ 2023-08-10T06:32:49,938 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.MapReduceBackupCopyJob(373): Doing SNAPSHOT_COPY 2023-08-10T06:32:49,963 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1003): inputFs=hdfs://localhost:33549, inputRoot=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:49,963 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1004): outputFs=DFS[DFSClient[clientName=DFSClient_hb_m_jenkins-hbase3.apache.org,42279,1691649119143_1795665932_16, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/.hbase-snapshot/.tmp/snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:49,963 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1009): Verify snapshot source, inputFs=hdfs://localhost:33549, inputRoot=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3, snapshotDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/snapshot_1691649167811_default_test-1691649135102. 2023-08-10T06:32:49,976 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1068): Copy Snapshot Manifest from hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/.hbase-snapshot/snapshot_1691649167811_default_test-1691649135102 to hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/.hbase-snapshot/.tmp/snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:50,393 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:50,393 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:50,394 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:50,394 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,307 INFO [master/jenkins-hbase3:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2023-08-10T06:32:51,308 INFO [master/jenkins-hbase3:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2023-08-10T06:32:51,407 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop-2880074060960275126.jar 2023-08-10T06:32:51,408 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,408 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,408 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,409 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,409 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,410 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,410 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,410 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,411 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:51,411 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.4/hbase-shaded-miscellaneous-4.1.4.jar 2023-08-10T06:32:51,412 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.4/hbase-shaded-gson-4.1.4.jar 2023-08-10T06:32:51,413 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.4/hbase-shaded-protobuf-4.1.4.jar 2023-08-10T06:32:51,413 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.4/hbase-shaded-netty-4.1.4.jar 2023-08-10T06:32:51,413 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.4/hbase-unsafe-4.1.4.jar 2023-08-10T06:32:51,414 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.5.7/zookeeper-3.5.7.jar 2023-08-10T06:32:51,414 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2023-08-10T06:32:51,415 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2023-08-10T06:32:51,415 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2023-08-10T06:32:51,416 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2023-08-10T06:32:51,416 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2023-08-10T06:32:51,416 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2023-08-10T06:32:51,417 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/2.10.2/hadoop-common-2.10.2.jar 2023-08-10T06:32:51,418 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/2.10.2/hadoop-common-2.10.2.jar 2023-08-10T06:32:51,418 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.10.2/hadoop-mapreduce-client-core-2.10.2.jar 2023-08-10T06:32:51,419 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/2.10.2/hadoop-common-2.10.2.jar 2023-08-10T06:32:51,419 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/2.10.2/hadoop-common-2.10.2.jar 2023-08-10T06:32:51,419 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.10.2/hadoop-mapreduce-client-core-2.10.2.jar 2023-08-10T06:32:51,420 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.10.2/hadoop-mapreduce-client-core-2.10.2.jar 2023-08-10T06:32:51,451 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] mapreduce.JobResourceUploader(142): Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this. 2023-08-10T06:32:51,473 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(581): Loading Snapshot 'snapshot_1691649167811_default_test-1691649135102' hfile list 2023-08-10T06:32:51,477 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(674): export split=0 size=8.2 K 2023-08-10T06:32:51,564 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] fs.FileUtil(888): Command 'ln -s /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.local.dir/1691649171522/libjars /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/*' failed 1 with: ln: failed to create symbolic link '/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/*': No such file or directory 2023-08-10T06:32:51,564 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] mapred.LocalDistributedCacheManager(202): Failed to create symlink: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/mapreduce.cluster.local.dir/1691649171522/libjars <- /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/libjars/* 2023-08-10T06:32:51,619 INFO [LocalJobRunner Map Task Executor #0 {}] snapshot.ExportSnapshot$ExportMapper(228): Using bufferSize=128 M 2023-08-10T06:32:51,691 INFO [LocalJobRunner Map Task Executor #0 {}] snapshot.ExportSnapshot$ExportMapper(454): copy completed for input=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de output=hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:51,692 INFO [LocalJobRunner Map Task Executor #0 {}] snapshot.ExportSnapshot$ExportMapper(456): size=8396 (8.2 K) time=0sec 8.007M/sec 2023-08-10T06:32:52,602 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1125): Finalize the Snapshot Export 2023-08-10T06:32:52,603 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1136): Verify snapshot integrity 2023-08-10T06:32:52,610 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] snapshot.ExportSnapshot(1140): Export Completed: snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:52,610 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.FullTableBackupClient(121): Snapshot copy snapshot_1691649167811_default_test-1691649135102 finished. 2023-08-10T06:32:52,610 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(1014): test-1691649135102 2023-08-10T06:32:52,617 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManager(277): Getting the direct ancestors of the current backup backup_1691649163473 2023-08-10T06:32:52,617 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManager(283): Current backup is a full backup, no direct ancestor for it. 2023-08-10T06:32:53,022 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupManifest(484): Manifest file stored to hdfs://localhost:44455/backupUT/backup_1691649163473/.backup.manifest 2023-08-10T06:32:53,022 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(358): Backup backup_1691649163473 finished: type=FULL,tablelist=test-1691649135102,targetRootDir=hdfs://localhost:44455/backupUT,startts=1691649167620,completets=1691649172617,bytescopied=0 2023-08-10T06:32:53,022 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(138): Trying to delete snapshot for full backup. 2023-08-10T06:32:53,023 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(143): Trying to delete snapshot: snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:53,023 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(759): Client=jenkins//172.31.12.81 delete name: "snapshot_1691649167811_default_test-1691649135102" 2023-08-10T06:32:53,026 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(378): Deleting snapshot: snapshot_1691649167811_default_test-1691649135102 2023-08-10T06:32:53,027 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(148): Deleting the snapshot snapshot_1691649167811_default_test-1691649135102 for backup backup_1691649163473 succeeded. 2023-08-10T06:32:53,027 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(1552): Deleting snapshot_backup_system from the system 2023-08-10T06:32:53,032 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(759): Client=jenkins//172.31.12.81 delete name: "snapshot_backup_system" 2023-08-10T06:32:53,034 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] snapshot.SnapshotManager(378): Deleting snapshot: snapshot_backup_system 2023-08-10T06:32:53,035 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(1557): Done deleting backup system table snapshot 2023-08-10T06:32:53,036 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] impl.BackupSystemTable(637): Finish backup exclusive operation 2023-08-10T06:32:53,042 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] impl.TableBackupClient(376): Backup backup_1691649163473 completed. 2023-08-10T06:32:53,043 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:32:53,043 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x0d3a42bc to 127.0.0.1:52413 2023-08-10T06:32:53,043 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:32:53,043 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] backup.TestRemoteRestore(94): backup complete 2023-08-10T06:32:53,054 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.log.dir so I do NOT create it in target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d 2023-08-10T06:32:53,054 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.tmp.dir so I do NOT create it in target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d 2023-08-10T06:32:53,054 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(455): hadoop.tmp.dir property value differs in configuration and system: Configuration=/tmp/hadoop-jenkins while System=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.tmp.dir Erasing configuration value by system value. 2023-08-10T06:32:53,054 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(348): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d 2023-08-10T06:32:53,055 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc, deleteOnExit=true 2023-08-10T06:32:53,055 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/test.cache.data in system properties and HBase conf 2023-08-10T06:32:53,055 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/hadoop.tmp.dir in system properties and HBase conf 2023-08-10T06:32:53,055 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/hadoop.log.dir in system properties and HBase conf 2023-08-10T06:32:53,055 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/mapreduce.cluster.local.dir in system properties and HBase conf 2023-08-10T06:32:53,055 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/mapreduce.cluster.temp.dir in system properties and HBase conf 2023-08-10T06:32:53,055 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2023-08-10T06:32:53,055 DEBUG [Listener at jenkins-hbase3.apache.org/41665 {}] fs.HFileSystem(308): The file system is not a DistributedFileSystem. Skipping on block location reordering 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.nodemanager.log-dirs in system properties and HBase conf 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:32:53,056 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2023-08-10T06:32:53,057 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/nfs.dump.dir in system properties and HBase conf 2023-08-10T06:32:53,057 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir in system properties and HBase conf 2023-08-10T06:32:53,057 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:32:53,057 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2023-08-10T06:32:53,057 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/fs.s3a.committer.staging.tmp.path in system properties and HBase conf Formatting using clusterid: testClusterID 2023-08-10T06:32:53,069 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] conf.Configuration(1733): No unit for dfs.heartbeat.interval(3) assuming SECONDS 2023-08-10T06:32:53,069 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] conf.Configuration(1733): No unit for dfs.namenode.safemode.extension(0) assuming MILLISECONDS 2023-08-10T06:32:53,119 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:53,121 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:53,128 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/hdfs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir/Jetty_localhost_46353_hdfs____4ce321/webapp 2023-08-10T06:32:53,229 INFO [Listener at jenkins-hbase3.apache.org/41665 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:46353 2023-08-10T06:32:53,241 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] conf.Configuration(1733): No unit for dfs.heartbeat.interval(3) assuming SECONDS 2023-08-10T06:32:53,241 WARN [Listener at jenkins-hbase3.apache.org/41665 {}] conf.Configuration(1733): No unit for dfs.namenode.safemode.extension(0) assuming MILLISECONDS 2023-08-10T06:32:53,298 WARN [Listener at localhost/40033 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:32:53,314 WARN [Listener at localhost/40033 {}] conf.Configuration(1733): No unit for dfs.datanode.outliers.report.interval(1800000) assuming MILLISECONDS 2023-08-10T06:32:53,316 WARN [Listener at localhost/40033 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:53,317 INFO [Listener at localhost/40033 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:53,323 INFO [Listener at localhost/40033 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/datanode to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir/Jetty_localhost_38607_datanode____.w6xhi9/webapp 2023-08-10T06:32:53,420 INFO [Listener at localhost/40033 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:38607 2023-08-10T06:32:53,436 WARN [Listener at localhost/43257 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:32:53,452 WARN [Listener at localhost/43257 {}] conf.Configuration(1733): No unit for dfs.datanode.outliers.report.interval(1800000) assuming MILLISECONDS 2023-08-10T06:32:53,453 WARN [Listener at localhost/43257 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:53,455 INFO [Listener at localhost/43257 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:53,459 INFO [Listener at localhost/43257 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/datanode to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir/Jetty_localhost_32801_datanode____qexkrl/webapp 2023-08-10T06:32:53,514 WARN [Thread-1148 {}] conf.Configuration(1733): No unit for dfs.datanode.lock-reporting-threshold-ms(300) assuming MILLISECONDS 2023-08-10T06:32:53,544 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x334e34b013b4f944: Processing first storage report for DS-4530c01e-9254-42f5-b845-b99a718c3dd7 from datanode 3755aa3f-4c85-4fae-bf9c-862f69abca10 2023-08-10T06:32:53,544 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x334e34b013b4f944: from storage DS-4530c01e-9254-42f5-b845-b99a718c3dd7 node DatanodeRegistration(127.0.0.1:46841, datanodeUuid=3755aa3f-4c85-4fae-bf9c-862f69abca10, infoPort=36197, infoSecurePort=0, ipcPort=43257, storageInfo=lv=-57;cid=testClusterID;nsid=1351540680;c=1691649173071), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2023-08-10T06:32:53,544 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x334e34b013b4f944: Processing first storage report for DS-973be953-ff4f-4ef8-a274-ac9e18f4e275 from datanode 3755aa3f-4c85-4fae-bf9c-862f69abca10 2023-08-10T06:32:53,544 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x334e34b013b4f944: from storage DS-973be953-ff4f-4ef8-a274-ac9e18f4e275 node DatanodeRegistration(127.0.0.1:46841, datanodeUuid=3755aa3f-4c85-4fae-bf9c-862f69abca10, infoPort=36197, infoSecurePort=0, ipcPort=43257, storageInfo=lv=-57;cid=testClusterID;nsid=1351540680;c=1691649173071), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:53,574 INFO [Listener at localhost/43257 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32801 2023-08-10T06:32:53,591 WARN [Listener at localhost/45831 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:32:53,626 WARN [Listener at localhost/45831 {}] conf.Configuration(1733): No unit for dfs.datanode.outliers.report.interval(1800000) assuming MILLISECONDS 2023-08-10T06:32:53,628 WARN [Listener at localhost/45831 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:53,629 INFO [Listener at localhost/45831 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:53,638 INFO [Listener at localhost/45831 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/2.10.2/hadoop-hdfs-2.10.2-tests.jar!/webapps/datanode to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir/Jetty_localhost_41911_datanode____2s2ctf/webapp 2023-08-10T06:32:53,682 WARN [Thread-1183 {}] conf.Configuration(1733): No unit for dfs.datanode.lock-reporting-threshold-ms(300) assuming MILLISECONDS 2023-08-10T06:32:53,715 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x889cd642e841a145: Processing first storage report for DS-7de55a91-d7ee-42a2-a4a6-661d59e7d171 from datanode 9b7244fa-ab72-4491-bc61-65d6f5373f6f 2023-08-10T06:32:53,715 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x889cd642e841a145: from storage DS-7de55a91-d7ee-42a2-a4a6-661d59e7d171 node DatanodeRegistration(127.0.0.1:37005, datanodeUuid=9b7244fa-ab72-4491-bc61-65d6f5373f6f, infoPort=45223, infoSecurePort=0, ipcPort=45831, storageInfo=lv=-57;cid=testClusterID;nsid=1351540680;c=1691649173071), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:53,715 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x889cd642e841a145: Processing first storage report for DS-dab7e4b6-0c69-449e-a4da-6c839e9b38b0 from datanode 9b7244fa-ab72-4491-bc61-65d6f5373f6f 2023-08-10T06:32:53,715 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x889cd642e841a145: from storage DS-dab7e4b6-0c69-449e-a4da-6c839e9b38b0 node DatanodeRegistration(127.0.0.1:37005, datanodeUuid=9b7244fa-ab72-4491-bc61-65d6f5373f6f, infoPort=45223, infoSecurePort=0, ipcPort=45831, storageInfo=lv=-57;cid=testClusterID;nsid=1351540680;c=1691649173071), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:53,750 INFO [Listener at localhost/45831 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41911 2023-08-10T06:32:53,771 WARN [Listener at localhost/40215 {}] common.MetricsLoggerTask(153): Metrics logging will not be async since the logger is not log4j 2023-08-10T06:32:53,841 WARN [Thread-1217 {}] conf.Configuration(1733): No unit for dfs.datanode.lock-reporting-threshold-ms(300) assuming MILLISECONDS 2023-08-10T06:32:53,867 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x45d04259b3f9d445: Processing first storage report for DS-af57fbfe-ec82-42e8-b122-13fe57b143db from datanode ca86ccdf-0fa2-46a1-957d-8e975b17032a 2023-08-10T06:32:53,867 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x45d04259b3f9d445: from storage DS-af57fbfe-ec82-42e8-b122-13fe57b143db node DatanodeRegistration(127.0.0.1:39019, datanodeUuid=ca86ccdf-0fa2-46a1-957d-8e975b17032a, infoPort=45277, infoSecurePort=0, ipcPort=40215, storageInfo=lv=-57;cid=testClusterID;nsid=1351540680;c=1691649173071), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:53,867 INFO [Block report processor {}] blockmanagement.BlockManager(2202): BLOCK* processReport 0x45d04259b3f9d445: Processing first storage report for DS-306c97f3-dcb2-4c58-bb4e-50db94b3a77f from datanode ca86ccdf-0fa2-46a1-957d-8e975b17032a 2023-08-10T06:32:53,868 INFO [Block report processor {}] blockmanagement.BlockManager(2228): BLOCK* processReport 0x45d04259b3f9d445: from storage DS-306c97f3-dcb2-4c58-bb4e-50db94b3a77f node DatanodeRegistration(127.0.0.1:39019, datanodeUuid=ca86ccdf-0fa2-46a1-957d-8e975b17032a, infoPort=45277, infoSecurePort=0, ipcPort=40215, storageInfo=lv=-57;cid=testClusterID;nsid=1351540680;c=1691649173071), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2023-08-10T06:32:53,882 DEBUG [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d 2023-08-10T06:32:53,882 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(2786): Starting mini mapreduce cluster... 2023-08-10T06:32:53,882 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/test.cache.data in system properties and HBase conf 2023-08-10T06:32:53,882 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/hadoop.tmp.dir in system properties and HBase conf 2023-08-10T06:32:53,882 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/hadoop.log.dir in system properties and HBase conf 2023-08-10T06:32:53,882 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/mapreduce.cluster.local.dir in system properties and HBase conf 2023-08-10T06:32:53,882 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/mapreduce.cluster.temp.dir in system properties and HBase conf 2023-08-10T06:32:53,882 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2023-08-10T06:32:53,882 INFO [Listener at localhost/40215 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:53,883 INFO [Listener at localhost/40215 {}] fs.HFileSystem(337): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.nodemanager.log-dirs in system properties and HBase conf 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:32:53,886 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2023-08-10T06:32:53,887 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/nfs.dump.dir in system properties and HBase conf 2023-08-10T06:32:53,887 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir in system properties and HBase conf 2023-08-10T06:32:53,887 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/dfs.journalnode.edits.dir in system properties and HBase conf 2023-08-10T06:32:53,887 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2023-08-10T06:32:53,887 INFO [Listener at localhost/40215 {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2023-08-10T06:32:54,345 WARN [Listener at localhost/40215 {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2023-08-10T06:32:54,703 WARN [Thread-1337 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:54,707 WARN [Thread-1337 {}] servlet.GuiceFilter(87): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2023-08-10T06:32:54,708 INFO [Thread-1337 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:54,714 INFO [Thread-1337 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/2.10.2/hadoop-yarn-common-2.10.2.jar!/webapps/jobhistory to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir/Jetty_jenkins.hbase3_apache_org_43753_jobhistory____dc8mio/webapp 2023-08-10T06:32:54,763 INFO [Thread-1337 {}] container.GuiceComponentProviderFactory(107): Registering org.apache.hadoop.mapreduce.v2.hs.webapp.HsWebServices as a root resource class 2023-08-10T06:32:54,763 INFO [Thread-1337 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.mapreduce.v2.hs.webapp.JAXBContextResolver as a provider class 2023-08-10T06:32:54,763 INFO [Thread-1337 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2023-08-10T06:32:54,763 INFO [Thread-1337 {}] application.WebApplicationImpl(791): Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM' 2023-08-10T06:32:54,770 INFO [Thread-1337 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.mapreduce.v2.hs.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:54,815 INFO [Thread-1337 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:54,865 INFO [Thread-1337 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.mapreduce.v2.hs.webapp.HsWebServices to GuiceManagedComponentProvider with the scope "PerRequest" 2023-08-10T06:32:54,871 INFO [Thread-1337 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:43753 2023-08-10T06:32:56,193 WARN [Listener at jenkins-hbase3.apache.org/32987 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:56,201 WARN [Listener at jenkins-hbase3.apache.org/32987 {}] servlet.GuiceFilter(87): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2023-08-10T06:32:56,201 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:56,209 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/2.10.2/hadoop-yarn-common-2.10.2.jar!/webapps/cluster to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir/Jetty_jenkins.hbase3_apache_org_33631_cluster____.nyu7td/webapp 2023-08-10T06:32:56,261 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver as a provider class 2023-08-10T06:32:56,261 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] container.GuiceComponentProviderFactory(107): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices as a root resource class 2023-08-10T06:32:56,261 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2023-08-10T06:32:56,261 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] application.WebApplicationImpl(791): Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM' 2023-08-10T06:32:56,268 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:56,348 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:56,409 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:56,415 INFO [Listener at jenkins-hbase3.apache.org/32987 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:33631 2023-08-10T06:32:56,664 WARN [Listener at jenkins-hbase3.apache.org/35967 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:56,671 WARN [Listener at jenkins-hbase3.apache.org/35967 {}] servlet.GuiceFilter(87): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2023-08-10T06:32:56,671 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:56,678 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/2.10.2/hadoop-yarn-common-2.10.2.jar!/webapps/node to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir/Jetty_jenkins.hbase3_apache_org_39979_node____8dfbo4/webapp 2023-08-10T06:32:56,727 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] container.GuiceComponentProviderFactory(107): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2023-08-10T06:32:56,727 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2023-08-10T06:32:56,727 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2023-08-10T06:32:56,727 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] application.WebApplicationImpl(791): Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM' 2023-08-10T06:32:56,734 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:56,748 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:56,801 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:56,808 INFO [Listener at jenkins-hbase3.apache.org/35967 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:39979 2023-08-10T06:32:56,894 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] http.HttpRequestLog(97): Jetty request log can only be enabled using Log4j 2023-08-10T06:32:56,899 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] servlet.GuiceFilter(87): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2023-08-10T06:32:56,900 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): jetty-6.1.26 2023-08-10T06:32:56,907 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Extract jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/2.10.2/hadoop-yarn-common-2.10.2.jar!/webapps/node to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/java.io.tmpdir/Jetty_jenkins.hbase3_apache_org_37093_node____i8fy4l/webapp 2023-08-10T06:32:56,956 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] container.GuiceComponentProviderFactory(107): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2023-08-10T06:32:56,956 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2023-08-10T06:32:56,956 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] container.GuiceComponentProviderFactory(104): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2023-08-10T06:32:56,956 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] application.WebApplicationImpl(791): Initiating Jersey application, version 'Jersey: 1.9 09/02/2011 11:17 AM' 2023-08-10T06:32:56,963 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:56,977 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:57,040 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] container.GuiceComponentProviderFactory(159): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2023-08-10T06:32:57,047 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:37093 2023-08-10T06:32:57,051 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.HBaseTestingUtility(2821): Mini mapreduce cluster started 2023-08-10T06:32:57,051 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x6fbf2c4a to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:57,058 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2a6d4c4d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:57,060 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:57,065 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:44660, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:57,068 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] impl.RestoreTablesClient(101): HBase table table1 does not exist. It will be created during restore process 2023-08-10T06:32:57,069 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] impl.BackupManifest(413): Loading manifest from: hdfs://localhost:44455/backupUT/backup_1691649163473/.backup.manifest 2023-08-10T06:32:57,072 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] impl.BackupManifest(443): Loaded manifest instance from manifest file: /backupUT/backup_1691649163473/.backup.manifest 2023-08-10T06:32:57,073 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] impl.BackupManifest(413): Loading manifest from: hdfs://localhost:44455/backupUT/backup_1691649163473/.backup.manifest 2023-08-10T06:32:57,075 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] impl.BackupManifest(443): Loaded manifest instance from manifest file: /backupUT/backup_1691649163473/.backup.manifest 2023-08-10T06:32:57,075 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] impl.RestoreTablesClient(148): Restoring 'test-1691649135102' to 'table1' from full backup image hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102 2023-08-10T06:32:57,084 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] util.RestoreTool(487): Creating target table 'table1' 2023-08-10T06:32:57,084 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] util.RestoreTool(391): Parsing region dir: hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:57,085 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] util.RestoreTool(421): Parsing family dir [hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f in region [hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3] 2023-08-10T06:32:57,089 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] util.RestoreTool(442): Trying to figure out region boundaries hfile=hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/f/317474dc8ece4aa0a1b9b2a97a3c57de first=row0 last=row98 2023-08-10T06:32:57,089 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:32:57,090 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:57,094 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:39738, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:32:57,096 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$4(2331): Client=jenkins//172.31.12.81 create 'table1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2023-08-10T06:32:57,098 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=47, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=table1 2023-08-10T06:32:57,100 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=47, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_PRE_OPERATION 2023-08-10T06:32:57,101 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(708): Client=jenkins//172.31.12.81 procedure request for creating table: namespace: "default" qualifier: "table1" procId is: 47 2023-08-10T06:32:57,101 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=47, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2023-08-10T06:32:57,101 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=47 2023-08-10T06:32:57,203 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=47 2023-08-10T06:32:57,404 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=47 2023-08-10T06:32:57,509 INFO [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(7075): creating {ENCODED => 95c47c287fe0556429d6da3716d1d43d, NAME => 'table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='table1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'f', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3 2023-08-10T06:32:57,705 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=47 2023-08-10T06:32:57,916 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(888): Instantiated table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:57,916 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1668): Closing 95c47c287fe0556429d6da3716d1d43d, disabling compactions & flushes 2023-08-10T06:32:57,916 INFO [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1690): Closing region table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:32:57,916 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:32:57,916 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1778): Acquired close lock on table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. after waiting 0 ms 2023-08-10T06:32:57,916 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1788): Updates disabled for region table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:32:57,916 INFO [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1909): Closed table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:32:57,916 DEBUG [RegionOpenAndInit-table1-pool-0 {}] regionserver.HRegion(1622): Region close journal for 95c47c287fe0556429d6da3716d1d43d: 2023-08-10T06:32:57,917 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=47, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_ADD_TO_META 2023-08-10T06:32:57,917 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d.","families":{"info":[{"qualifier":"regioninfo","vlen":40,"tag":[],"timestamp":"1691649177917"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1691649177917"}]},"ts":"1691649177917"} 2023-08-10T06:32:57,919 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2023-08-10T06:32:57,920 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(81): pid=47, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_ASSIGN_REGIONS 2023-08-10T06:32:57,920 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649177920"}]},"ts":"1691649177920"} 2023-08-10T06:32:57,922 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=table1, state=ENABLING in hbase:meta 2023-08-10T06:32:57,925 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=48, ppid=47, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=table1, region=95c47c287fe0556429d6da3716d1d43d, ASSIGN}] 2023-08-10T06:32:57,928 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=48, ppid=47, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=table1, region=95c47c287fe0556429d6da3716d1d43d, ASSIGN 2023-08-10T06:32:57,929 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(262): Starting pid=48, ppid=47, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=table1, region=95c47c287fe0556429d6da3716d1d43d, ASSIGN; state=OFFLINE, location=jenkins-hbase3.apache.org,34991,1691649122119; forceNewPlan=false, retain=false 2023-08-10T06:32:58,082 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=48 updating hbase:meta row=95c47c287fe0556429d6da3716d1d43d, regionState=OPENING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:58,084 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=49, ppid=48, state=RUNNABLE; OpenRegionProcedure 95c47c287fe0556429d6da3716d1d43d, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:32:58,206 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=47 2023-08-10T06:32:58,239 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] handler.AssignRegionHandler(130): Open table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:32:58,239 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(7254): Opening region: {ENCODED => 95c47c287fe0556429d6da3716d1d43d, NAME => 'table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d.', STARTKEY => '', ENDKEY => ''} 2023-08-10T06:32:58,239 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.backup.BackupObserver loaded, priority=536870911. 2023-08-10T06:32:58,239 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table table1 95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:32:58,240 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(888): Instantiated table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2023-08-10T06:32:58,240 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(7296): checking encryption for 95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:32:58,240 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(7299): checking classloading for 95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:32:58,241 INFO [StoreOpener-95c47c287fe0556429d6da3716d1d43d-1 {}] regionserver.HStore(397): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family f of region 95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:32:58,242 INFO [StoreOpener-95c47c287fe0556429d6da3716d1d43d-1 {}] compactions.CompactionConfiguration(173): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 95c47c287fe0556429d6da3716d1d43d columnFamilyName f 2023-08-10T06:32:58,243 DEBUG [StoreOpener-95c47c287fe0556429d6da3716d1d43d-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2023-08-10T06:32:58,243 INFO [StoreOpener-95c47c287fe0556429d6da3716d1d43d-1 {}] regionserver.HStore(324): Store=95c47c287fe0556429d6da3716d1d43d/f, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2023-08-10T06:32:58,244 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:32:58,244 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(5270): Found 0 recovered edits file(s) under hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:32:58,247 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(1076): writing seq id for 95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:32:58,248 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2023-08-10T06:32:58,249 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(1093): Opened 95c47c287fe0556429d6da3716d1d43d; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=11475705440, jitterRate=0.06875835359096527}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2023-08-10T06:32:58,249 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegion(992): Region open journal for 95c47c287fe0556429d6da3716d1d43d: 2023-08-10T06:32:58,250 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegionServer(2552): Post open deploy tasks for table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d., pid=49, masterSystemTime=1691649178235 2023-08-10T06:32:58,251 DEBUG [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] regionserver.HRegionServer(2579): Finished post open deploy task for table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:32:58,252 INFO [RS_OPEN_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_OPEN_REGION, pid=49}] handler.AssignRegionHandler(158): Opened table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:32:58,252 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=48 updating hbase:meta row=95c47c287fe0556429d6da3716d1d43d, regionState=OPEN, openSeqNum=2, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:32:58,255 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=49, resume processing ppid=48 2023-08-10T06:32:58,255 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1414): Finished pid=49, ppid=48, state=SUCCESS; OpenRegionProcedure 95c47c287fe0556429d6da3716d1d43d, server=jenkins-hbase3.apache.org,34991,1691649122119 in 169 msec 2023-08-10T06:32:58,257 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=48, resume processing ppid=47 2023-08-10T06:32:58,257 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=48, ppid=47, state=SUCCESS; TransitRegionStateProcedure table=table1, region=95c47c287fe0556429d6da3716d1d43d, ASSIGN in 330 msec 2023-08-10T06:32:58,257 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=47, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2023-08-10T06:32:58,257 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649178257"}]},"ts":"1691649178257"} 2023-08-10T06:32:58,259 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=table1, state=ENABLED in hbase:meta 2023-08-10T06:32:58,262 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(81): pid=47, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=table1 execute state=CREATE_TABLE_POST_OPERATION 2023-08-10T06:32:58,263 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=47, state=SUCCESS; CreateTableProcedure table=table1 in 1.1660 sec 2023-08-10T06:32:58,430 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'table1' 2023-08-10T06:32:59,207 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=47 2023-08-10T06:32:59,207 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] client.HBaseAdmin$TableFuture(3590): Operation: CREATE, Table Name: default:table1, procId: 47 completed 2023-08-10T06:32:59,216 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.MapReduceRestoreJob(64): Restore full backup from directory hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3 from hbase tables test-1691649135102 to tables table1 2023-08-10T06:32:59,216 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.MapReduceRestoreJob(72): Restore test-1691649135102 into table1 2023-08-10T06:32:59,217 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.MapReduceHFileSplitterJob(107): add incremental job :hdfs://localhost:33549/backupUT/bulk_output-default-table1-1691649179216 from hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/archive/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:32:59,218 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x16114814 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:32:59,224 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@44f0d8b6, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:32:59,224 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:32:59,226 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:39750, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:32:59,227 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.HFileOutputFormat2(672): bulkload locality sensitive enabled 2023-08-10T06:32:59,227 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.HFileOutputFormat2(535): Looking up current regions for table table1 2023-08-10T06:32:59,232 DEBUG [hconnection-0x5dd9be16-shared-pool-0 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:32:59,234 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:44676, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:32:59,235 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.HFileOutputFormat2(544): SplitPoint startkey for table1: 2023-08-10T06:32:59,235 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.HFileOutputFormat2(694): Configuring 1 reduce partitions to match current region count for all tables 2023-08-10T06:32:59,235 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.HFileOutputFormat2(559): Writing partition information to /user/jenkins/hbase-staging/partitions_48b7293b-2c55-4b50-b067-3506b2b8ffa8 2023-08-10T06:32:59,245 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:59,245 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:59,246 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:59,246 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.6.0-SNAPSHOT.jar 2023-08-10T06:32:59,308 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_table1 2023-08-10T06:32:59,308 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_table1 Metrics about Tables on a single HBase RegionServer 2023-08-10T06:33:00,208 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/hadoop-2580962811300084455.jar 2023-08-10T06:33:00,208 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,208 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,209 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,209 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,210 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,210 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,210 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,211 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,211 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,212 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.4/hbase-shaded-miscellaneous-4.1.4.jar 2023-08-10T06:33:00,213 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.4/hbase-shaded-gson-4.1.4.jar 2023-08-10T06:33:00,213 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.4/hbase-shaded-protobuf-4.1.4.jar 2023-08-10T06:33:00,213 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.4/hbase-shaded-netty-4.1.4.jar 2023-08-10T06:33:00,214 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.4/hbase-unsafe-4.1.4.jar 2023-08-10T06:33:00,214 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.5.7/zookeeper-3.5.7.jar 2023-08-10T06:33:00,214 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2023-08-10T06:33:00,215 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2023-08-10T06:33:00,215 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2023-08-10T06:33:00,216 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2023-08-10T06:33:00,216 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2023-08-10T06:33:00,216 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2023-08-10T06:33:00,217 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.io.ImmutableBytesWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,217 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.util.MapReduceExtendedCell, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,218 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.HFileInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,218 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.io.ImmutableBytesWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,218 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.util.MapReduceExtendedCell, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,219 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/hbase-mapreduce-2.6.0-SNAPSHOT.jar 2023-08-10T06:33:00,219 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.10.2/hadoop-mapreduce-client-core-2.10.2.jar 2023-08-10T06:33:00,219 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.HFileOutputFormat2(714): Incremental output configured for tables: table1 2023-08-10T06:33:00,220 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:33:00,220 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x16114814 to 127.0.0.1:52413 2023-08-10T06:33:00,220 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:00,220 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.MapReduceHFileSplitterJob(119): success configuring load incremental job 2023-08-10T06:33:00,220 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(796): The addDependencyJars(Configuration, Class...) method has been deprecated since it is easy to use incorrectly. Most users should rely on addDependencyJars(Job) instead. See HBASE-8386 for more details. 2023-08-10T06:33:00,221 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.TableMapReduceUtil(868): For class org.apache.hbase.thirdparty.com.google.common.base.Preconditions, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.4/hbase-shaded-miscellaneous-4.1.4.jar 2023-08-10T06:33:00,415 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.JobResourceUploader(142): Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this. 2023-08-10T06:33:00,790 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.JobResourceUploader(470): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2023-08-10T06:33:01,120 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.LeafQueue(815): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2023-08-10T06:33:01,120 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.LeafQueue(851): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2023-08-10T06:33:01,980 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2001): Auth successful for appattempt_1691649176425_0001_000001 (auth:SIMPLE) 2023-08-10T06:33:03,744 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2023-08-10T06:33:05,540 DEBUG [master/jenkins-hbase3:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 333dde743887b960947ea82e4c5cfa5e changed from -1.0 to 0.0, refreshing cache 2023-08-10T06:33:06,801 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2001): Auth successful for appattempt_1691649176425_0001_000001 (auth:SIMPLE) 2023-08-10T06:33:09,059 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2001): Auth successful for appattempt_1691649176425_0001_000001 (auth:SIMPLE) 2023-08-10T06:33:09,592 DEBUG [master/jenkins-hbase3:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2023-08-10T06:33:09,592 DEBUG [master/jenkins-hbase3:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 86e6ed14dc5c14940dff214bf60e7e35 changed from -1.0 to 0.0, refreshing cache 2023-08-10T06:33:14,954 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2001): Auth successful for appattempt_1691649176425_0001_000001 (auth:SIMPLE) 2023-08-10T06:33:18,101 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:55896, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:33:18,508 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2001): Auth successful for appattempt_1691649176425_0001_000001 (auth:SIMPLE) 2023-08-10T06:33:18,527 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(305): Exit code from container container_1691649176425_0001_01_000003 is : 143 2023-08-10T06:33:20,490 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.MapReduceRestoreJob(89): Restoring HFiles from directory hdfs://localhost:33549/backupUT/bulk_output-default-table1-1691649179216 2023-08-10T06:33:20,490 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x6f1a7481 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:33:20,500 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1126aee5, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:33:20,501 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:33:20,504 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:40586, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:33:20,514 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] tool.LoadIncrementalHFiles(1067): Skipping non-directory hdfs://localhost:33549/backupUT/bulk_output-default-table1-1691649179216/_SUCCESS 2023-08-10T06:33:20,519 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:33:20,522 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:49490, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:33:20,539 DEBUG [LoadIncrementalHFiles-0 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: f229f00d1cd34267a5078563c1275ac1 2023-08-10T06:33:20,539 INFO [LoadIncrementalHFiles-0 {}] tool.LoadIncrementalHFiles(794): Trying to load hfile=hdfs://localhost:33549/backupUT/bulk_output-default-table1-1691649179216/f/f229f00d1cd34267a5078563c1275ac1 first=Optional[row0] last=Optional[row98] 2023-08-10T06:33:20,540 DEBUG [LoadIncrementalHFiles-0 {}] tool.LoadIncrementalHFiles$1(559): Going to connect to server region=table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d., hostname=jenkins-hbase3.apache.org,34991,1691649122119, seqNum=2 for row with hfile group [{f,hdfs://localhost:33549/backupUT/bulk_output-default-table1-1691649179216/f/f229f00d1cd34267a5078563c1275ac1}] 2023-08-10T06:33:20,550 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(611): Validating hfile at hdfs://localhost:33549/backupUT/bulk_output-default-table1-1691649179216/f/f229f00d1cd34267a5078563c1275ac1 for inclusion in 95c47c287fe0556429d6da3716d1d43d/f 2023-08-10T06:33:20,553 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(623): HFile bounds: first=row0 last=row98 2023-08-10T06:33:20,553 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(625): Region bounds: first= last= 2023-08-10T06:33:20,554 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: f229f00d1cd34267a5078563c1275ac1 2023-08-10T06:33:20,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HRegion(2520): Flush status journal for 95c47c287fe0556429d6da3716d1d43d: 2023-08-10T06:33:20,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.SecureBulkLoadManager$SecureBulkLoadListener(397): Moving hdfs://localhost:33549/backupUT/bulk_output-default-table1-1691649179216/f/f229f00d1cd34267a5078563c1275ac1 to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__mmpfmlg84hndd87u8kknupqi5p2tnsc1bb1h67id98t2id293a8diajh1pvis6tp/f/f229f00d1cd34267a5078563c1275ac1 2023-08-10T06:33:20,559 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__mmpfmlg84hndd87u8kknupqi5p2tnsc1bb1h67id98t2id293a8diajh1pvis6tp/f/f229f00d1cd34267a5078563c1275ac1 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/f/4eb5d1681f014d12a6050e4a330a1bf6_SeqId_4_ 2023-08-10T06:33:20,559 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x22aca485 to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:33:20,565 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7cd37125, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:33:20,565 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:33:20,567 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:49506, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2023-08-10T06:33:20,571 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:33:20,572 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:40596, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2023-08-10T06:33:20,586 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:33:20,586 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x22aca485 to 127.0.0.1:52413 2023-08-10T06:33:20,586 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:20,587 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(697): Loaded HFile hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__mmpfmlg84hndd87u8kknupqi5p2tnsc1bb1h67id98t2id293a8diajh1pvis6tp/f/f229f00d1cd34267a5078563c1275ac1 into 95c47c287fe0556429d6da3716d1d43d/f as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/f/4eb5d1681f014d12a6050e4a330a1bf6_SeqId_4_ - updating store file list. 2023-08-10T06:33:20,593 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(719): Loaded HFile hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/f/4eb5d1681f014d12a6050e4a330a1bf6_SeqId_4_ into 95c47c287fe0556429d6da3716d1d43d/f 2023-08-10T06:33:20,593 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.HStore(703): Successfully loaded hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__mmpfmlg84hndd87u8kknupqi5p2tnsc1bb1h67id98t2id293a8diajh1pvis6tp/f/f229f00d1cd34267a5078563c1275ac1 into 95c47c287fe0556429d6da3716d1d43d/f (new location: hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/f/4eb5d1681f014d12a6050e4a330a1bf6_SeqId_4_) 2023-08-10T06:33:20,594 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] regionserver.SecureBulkLoadManager$SecureBulkLoadListener(412): Bulk Load done for: hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/staging/jenkins__table1__mmpfmlg84hndd87u8kknupqi5p2tnsc1bb1h67id98t2id293a8diajh1pvis6tp/f/f229f00d1cd34267a5078563c1275ac1 2023-08-10T06:33:20,595 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] zookeeper.ReadOnlyZKClient(139): Connect 0x3d570a4f to 127.0.0.1:52413 with session timeout=90000ms, retries 30, retry interval 1000ms, keepAlive=60000ms 2023-08-10T06:33:20,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@49b9e70f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null 2023-08-10T06:33:20,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=MasterService, sasl=false 2023-08-10T06:33:20,603 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:49522, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=MasterService 2023-08-10T06:33:20,609 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.RpcConnection(127): Using SIMPLE authentication for service=ClientService, sasl=false 2023-08-10T06:33:20,612 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(459): Connection from 172.31.12.81:40610, version=2.6.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=ClientService 2023-08-10T06:33:20,632 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:33:20,632 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x3d570a4f to 127.0.0.1:52413 2023-08-10T06:33:20,632 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34991 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:20,636 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:33:20,639 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x6f1a7481 to 127.0.0.1:52413 2023-08-10T06:33:20,639 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:20,639 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] mapreduce.MapReduceRestoreJob(100): Restore Job finished:0 2023-08-10T06:33:20,639 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] impl.RestoreTablesClient(231): Restore includes the following image(s): 2023-08-10T06:33:20,639 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] impl.RestoreTablesClient(233): Backup: backup_1691649163473 hdfs://localhost:44455/backupUT/backup_1691649163473/default/test-1691649135102/ 2023-08-10T06:33:20,639 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] impl.RestoreTablesClient(238): restoreStage finished 2023-08-10T06:33:20,645 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] client.HBaseAdmin$18(963): Started disable of table1 2023-08-10T06:33:20,645 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$12(2663): Client=jenkins//172.31.12.81 disable table1 2023-08-10T06:33:20,646 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=50, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=table1 2023-08-10T06:33:20,649 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=50 2023-08-10T06:33:20,650 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649200649"}]},"ts":"1691649200649"} 2023-08-10T06:33:20,651 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=table1, state=DISABLING in hbase:meta 2023-08-10T06:33:20,652 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(293): Set table1 to state=DISABLING 2023-08-10T06:33:20,653 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=51, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=table1, region=95c47c287fe0556429d6da3716d1d43d, UNASSIGN}] 2023-08-10T06:33:20,656 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(727): Took xlock for pid=51, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=table1, region=95c47c287fe0556429d6da3716d1d43d, UNASSIGN 2023-08-10T06:33:20,656 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=51 updating hbase:meta row=95c47c287fe0556429d6da3716d1d43d, regionState=CLOSING, regionLocation=jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:33:20,658 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1685): Initialized subprocedures=[{pid=52, ppid=51, state=RUNNABLE; CloseRegionProcedure 95c47c287fe0556429d6da3716d1d43d, server=jenkins-hbase3.apache.org,34991,1691649122119}] 2023-08-10T06:33:20,750 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=50 2023-08-10T06:33:20,809 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] handler.UnassignRegionHandler(119): Close 95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:33:20,809 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] regionserver.HRegion(1668): Closing 95c47c287fe0556429d6da3716d1d43d, disabling compactions & flushes 2023-08-10T06:33:20,810 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] regionserver.HRegion(1690): Closing region table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:33:20,810 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] regionserver.HRegion(1711): Waiting without time limit for close lock on table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:33:20,810 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] regionserver.HRegion(1778): Acquired close lock on table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. after waiting 0 ms 2023-08-10T06:33:20,810 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] regionserver.HRegion(1788): Updates disabled for region table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:33:20,812 DEBUG [StoreFileCloser-95c47c287fe0556429d6da3716d1d43d-f-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 4eb5d1681f014d12a6050e4a330a1bf6_SeqId_4_ 2023-08-10T06:33:20,819 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/recovered.edits/6.seqid, newMaxSeqId=6, maxSeqId=1 2023-08-10T06:33:20,819 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:20,820 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] regionserver.HRegion(1909): Closed table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d. 2023-08-10T06:33:20,820 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] regionserver.HRegion(1622): Region close journal for 95c47c287fe0556429d6da3716d1d43d: 2023-08-10T06:33:20,822 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION, pid=52}] handler.UnassignRegionHandler(163): Closed 95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:33:20,822 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=51 updating hbase:meta row=95c47c287fe0556429d6da3716d1d43d, regionState=CLOSED 2023-08-10T06:33:20,825 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=52, resume processing ppid=51 2023-08-10T06:33:20,825 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1414): Finished pid=52, ppid=51, state=SUCCESS; CloseRegionProcedure 95c47c287fe0556429d6da3716d1d43d, server=jenkins-hbase3.apache.org,34991,1691649122119 in 166 msec 2023-08-10T06:33:20,827 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1828): Finished subprocedure pid=51, resume processing ppid=50 2023-08-10T06:33:20,827 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1414): Finished pid=51, ppid=50, state=SUCCESS; TransitRegionStateProcedure table=table1, region=95c47c287fe0556429d6da3716d1d43d, UNASSIGN in 172 msec 2023-08-10T06:33:20,827 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1691649200827"}]},"ts":"1691649200827"} 2023-08-10T06:33:20,828 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=table1, state=DISABLED in hbase:meta 2023-08-10T06:33:20,830 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(305): Set table1 to state=DISABLED 2023-08-10T06:33:20,831 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1414): Finished pid=50, state=SUCCESS; DisableTableProcedure table=table1 in 185 msec 2023-08-10T06:33:20,951 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=50 2023-08-10T06:33:20,951 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] client.HBaseAdmin$TableFuture(3590): Operation: DISABLE, Table Name: default:table1, procId: 50 completed 2023-08-10T06:33:20,952 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.HMaster$5(2447): Client=jenkins//172.31.12.81 delete table1 2023-08-10T06:33:20,953 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] procedure2.ProcedureExecutor(1033): Stored pid=53, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=table1 2023-08-10T06:33:20,955 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(98): Waiting for RIT for pid=53, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:33:20,956 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(110): Deleting regions from filesystem for pid=53, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:33:20,956 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=53 2023-08-10T06:33:20,958 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:33:20,960 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/f, FileablePath, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/recovered.edits] 2023-08-10T06:33:20,965 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/f/4eb5d1681f014d12a6050e4a330a1bf6_SeqId_4_ to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/archive/data/default/table1/95c47c287fe0556429d6da3716d1d43d/f/4eb5d1681f014d12a6050e4a330a1bf6_SeqId_4_ 2023-08-10T06:33:20,968 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d/recovered.edits/6.seqid to hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/archive/data/default/table1/95c47c287fe0556429d6da3716d1d43d/recovered.edits/6.seqid 2023-08-10T06:33:20,969 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/table1/95c47c287fe0556429d6da3716d1d43d 2023-08-10T06:33:20,969 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(293): Archived table1 regions 2023-08-10T06:33:20,971 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(115): Deleting regions from META for pid=53, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:33:20,973 WARN [PEWorker-2 {}] procedure.DeleteTableProcedure(350): Deleting some vestigial 1 rows of table1 from hbase:meta 2023-08-10T06:33:20,974 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(387): Removing 'table1' descriptor. 2023-08-10T06:33:20,975 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(121): Deleting assignment state for pid=53, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:33:20,976 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(377): Removing 'table1' from region states. 2023-08-10T06:33:20,976 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1691649200976"}]},"ts":"9223372036854775807"} 2023-08-10T06:33:20,977 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1808): Deleted 1 regions from META 2023-08-10T06:33:20,977 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 95c47c287fe0556429d6da3716d1d43d, NAME => 'table1,,1691649177095.95c47c287fe0556429d6da3716d1d43d.', STARTKEY => '', ENDKEY => ''}] 2023-08-10T06:33:20,977 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(381): Marking 'table1' as deleted. 2023-08-10T06:33:20,978 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"table1","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1691649200977"}]},"ts":"9223372036854775807"} 2023-08-10T06:33:20,979 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1678): Deleted table table1 state from META 2023-08-10T06:33:20,982 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(127): Finished pid=53, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=table1 2023-08-10T06:33:20,983 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1414): Finished pid=53, state=SUCCESS; DeleteTableProcedure table=table1 in 30 msec 2023-08-10T06:33:21,057 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42279 {}] master.MasterRpcServices(1288): Checking to see if procedure is done pid=53 2023-08-10T06:33:21,057 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] client.HBaseAdmin$TableFuture(3590): Operation: DELETE, Table Name: default:table1, procId: 53 completed 2023-08-10T06:33:21,113 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.ResourceChecker(175): after: backup.TestRemoteRestore#testFullRestoreRemoteWithAlternateRestoreOutputDir Thread=1306 (was 831) Potentially hanging thread: 569211375@qtp-1974749666-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:38607 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498) org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192) org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124) org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Potentially hanging thread: IPC Server handler 43 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 45533 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: Timer-67 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 28 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 44 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 7 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Thread[Thread-1338,5,FailOnTimeoutGroup] java.lang.Thread.sleep(Native Method) org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover.run(AbstractDelegationTokenSecretManager.java:694) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 18 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 357184866@qtp-1770429183-0 java.lang.Object.wait(Native Method) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626) Potentially hanging thread: New I/O worker #55 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 27 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 15 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 19 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: MutableQuantiles-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #48 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data6) java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:627) Potentially hanging thread: Timer-55 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: pool-366-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: pool-351-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-17 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 12 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-56 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 8 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 2 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: pool-358-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: java.util.concurrent.ThreadPoolExecutor$Worker@30f042ff[State = -1, empty queue] sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 11 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@47ac1f3 sun.nio.ch.ServerSocketChannelImpl.accept0(Native Method) sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:421) sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:249) sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:113) org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:145) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 14 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 8 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 47 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 9 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 47 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 32 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-49 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 4 on default port 43257 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 34 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 0 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-53 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data2) java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:627) Potentially hanging thread: IPC Server handler 7 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 34 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-50 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 22 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 3 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 0 on default port 40215 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: pool-362-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: region-location-3 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1081) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data3) java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:627) Potentially hanging thread: IPC Server handler 35 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: BP-590518898-172.31.12.81-1691649173071 heartbeating to localhost/127.0.0.1:40033 java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:715) org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:851) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 41 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 11 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #45 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data5/current sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 4 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 11 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 40 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-48 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: org.apache.hadoop.util.JvmPauseMonitor$Monitor@7c473a1a java.lang.Thread.sleep(Native Method) org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:192) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: pool-398-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_2129524288_16 at /127.0.0.1:36632 [Receiving block BP-2138422956-172.31.12.81-1691649115962:blk_1073741877_1053] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read1(BufferedInputStream.java:286) java.io.BufferedInputStream.read(BufferedInputStream.java:345) java.io.DataInputStream.read(DataInputStream.java:149) org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:209) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:211) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109) org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:528) org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:968) org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:877) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:166) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:103) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 37 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: pool-349-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 38 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #38 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 33 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@7c49b5e5 java.lang.Thread.sleep(Native Method) org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:3975) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 0 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data5) java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:627) Potentially hanging thread: IPC Server handler 2 on default port 45831 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 0 on default port 35967 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 23 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 8 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 14 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 141980145@qtp-539198283-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:39979 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498) org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192) org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124) org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Potentially hanging thread: IPC Server handler 1 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 6 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Thread-2087 java.io.FileInputStream.readBytes(Native Method) java.io.FileInputStream.read(FileInputStream.java:255) java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.io.BufferedInputStream.read(BufferedInputStream.java:345) sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:284) sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:326) sun.nio.cs.StreamDecoder.read(StreamDecoder.java:178) java.io.InputStreamReader.read(InputStreamReader.java:184) java.io.BufferedReader.fill(BufferedReader.java:161) java.io.BufferedReader.readLine(BufferedReader.java:324) java.io.BufferedReader.readLine(BufferedReader.java:389) org.apache.hadoop.util.Shell$1.run(Shell.java:955) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-6 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 8 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-14 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Timer-64 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 3 on default port 35967 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data5/current/BP-590518898-172.31.12.81-1691649173071 java.lang.Thread.sleep(Native Method) org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:205) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: pool-356-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: org.apache.hadoop.util.JvmPauseMonitor$Monitor@6170c7c9 java.lang.Thread.sleep(Native Method) org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:192) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data1/current/BP-590518898-172.31.12.81-1691649173071 java.lang.Thread.sleep(Native Method) org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:205) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 3 on default port 41523 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 40215 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: Timer-65 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 0 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 33 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: DeletionService #0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1081) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 3 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 24 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 2 on default port 35967 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #41 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #50 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 39 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 44 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 41523 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: PacketResponder: BP-2138422956-172.31.12.81-1691649115962:blk_1073741876_1052, type=LAST_IN_PIPELINE java.lang.Object.wait(Native Method) java.lang.Object.wait(Object.java:502) org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.waitForAckHead(BlockReceiver.java:1327) org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1399) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: BP-590518898-172.31.12.81-1691649173071 heartbeating to localhost/127.0.0.1:40033 java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:715) org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:851) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@96908aa java.lang.Thread.sleep(Native Method) org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:536) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #63 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 1 on default port 43257 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 12 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 686785239@qtp-1240153792-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:43753 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498) org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192) org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124) org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Potentially hanging thread: IPC Server handler 7 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 31 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 9 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 49 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-59 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 1 on default port 41523 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 31 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #61 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: 689212283@qtp-1934758495-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:33631 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498) org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192) org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124) org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Potentially hanging thread: PacketResponder: BP-2138422956-172.31.12.81-1691649115962:blk_1073741877_1053, type=LAST_IN_PIPELINE java.lang.Object.wait(Native Method) java.lang.Object.wait(Object.java:502) org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.waitForAckHead(BlockReceiver.java:1327) org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1399) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 21 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-47 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: java.util.concurrent.ThreadPoolExecutor$Worker@129168ab[State = -1, empty queue] sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: ContainersLauncher #0 java.io.FileInputStream.readBytes(Native Method) java.io.FileInputStream.read(FileInputStream.java:255) java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.io.BufferedInputStream.read(BufferedInputStream.java:345) sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:284) sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:326) sun.nio.cs.StreamDecoder.read(StreamDecoder.java:178) java.io.InputStreamReader.read(InputStreamReader.java:184) java.io.BufferedReader.fill(BufferedReader.java:161) java.io.BufferedReader.read1(BufferedReader.java:212) java.io.BufferedReader.read(BufferedReader.java:286) org.apache.hadoop.util.Shell$ShellCommandExecutor.parseExecResult(Shell.java:1229) org.apache.hadoop.util.Shell.runCommand(Shell.java:984) org.apache.hadoop.util.Shell.run(Shell.java:884) org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1216) org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:294) org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.launchContainer(ContainerLaunch.java:447) org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:298) org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:99) java.util.concurrent.FutureTask.run(FutureTask.java:266) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Thread-1380 java.lang.Thread.sleep(Native Method) org.apache.hadoop.yarn.server.resourcemanager.scheduler.activities.ActivitiesManager$1.run(ActivitiesManager.java:143) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: 424756010@qtp-1934758495-0 java.lang.Object.wait(Native Method) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626) Potentially hanging thread: IPC Server handler 16 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 48 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 12 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 32987 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 2 on default port 43257 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 5 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 3 on default port 40215 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 8 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 10 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #49 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 37 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 2 on default port 40033 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/jenkins-hbase3:0-2 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 0 on default port 45533 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data6/current/BP-590518898-172.31.12.81-1691649173071 java.lang.Thread.sleep(Native Method) org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:205) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 10 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-39 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 0 on default port 41523 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: DeletionService #0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1081) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server idle connection scanner for port 35967 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: New I/O worker #40 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 17 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 16 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Client (1886842325) connection to localhost/127.0.0.1:40033 from jenkins java.lang.Object.wait(Native Method) org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1053) org.apache.hadoop.ipc.Client$Connection.run(Client.java:1097) Potentially hanging thread: IPC Server handler 42 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 32 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: pool-368-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 28 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 39 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O server boss #51 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) org.jboss.netty.channel.socket.nio.NioServerBoss.select(NioServerBoss.java:163) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 32 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #42 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-15 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 40 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 16 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 35967 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 48 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 15 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 24 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_242875349_16 at /127.0.0.1:48036 [Waiting for operation #8] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read(BufferedInputStream.java:265) java.io.DataInputStream.readShort(DataInputStream.java:312) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:67) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:269) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Client (1886842325) connection to jenkins-hbase3.apache.org/172.31.12.81:33779 from jenkins java.lang.Object.wait(Native Method) org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1053) org.apache.hadoop.ipc.Client$Connection.run(Client.java:1097) Potentially hanging thread: Thread[Thread-1373,5,FailOnTimeoutGroup] java.lang.Thread.sleep(Native Method) org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover.run(AbstractDelegationTokenSecretManager.java:694) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 36 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 14 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 17 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-5 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 2 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 46 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 47 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #53 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data3/current/BP-590518898-172.31.12.81-1691649173071 java.lang.Thread.sleep(Native Method) org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:205) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 30 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 35 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: DeletionService #1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1081) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 25 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 22 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 3 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-7 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 40 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #35 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #58 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 38 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 2 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: java.util.concurrent.ThreadPoolExecutor$Worker@4bb8d555[State = -1, empty queue] sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #60 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data1) java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:627) Potentially hanging thread: IPC Server handler 17 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 23 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 32504424@qtp-1816563711-0 java.lang.Object.wait(Native Method) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626) Potentially hanging thread: Timer-54 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 9 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 13 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #37 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: region-location-2 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1081) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 19 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 5 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 7 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 49 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 27 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 34931 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 4 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 34 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 40033 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 5 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 2 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 41 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: pool-352-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 3 on default port 45831 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 29 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 45831 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 304471555@qtp-669675177-0 java.lang.Object.wait(Native Method) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626) Potentially hanging thread: org.apache.hadoop.util.JvmPauseMonitor$Monitor@7bf61038 java.lang.Thread.sleep(Native Method) org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:192) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data4) java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:627) Potentially hanging thread: IPC Server idle connection scanner for port 43257 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 49 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 3 on default port 43257 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 45 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #47 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #54 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: ApplicationMasterLauncher #0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: org.apache.hadoop.util.JvmPauseMonitor$Monitor@30574a14 java.lang.Thread.sleep(Native Method) org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:192) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 19 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 30 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-51 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-9 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Listener at jenkins-hbase3.apache.org/41523 java.lang.Thread.dumpThreads(Native Method) java.lang.Thread.getAllStackTraces(Thread.java:1615) org.apache.hadoop.hbase.ResourceCheckerJUnitListener$ThreadResourceAnalyzer.getVal(ResourceCheckerJUnitListener.java:49) org.apache.hadoop.hbase.ResourceChecker.fill(ResourceChecker.java:110) org.apache.hadoop.hbase.ResourceChecker.fillEndings(ResourceChecker.java:104) org.apache.hadoop.hbase.ResourceChecker.end(ResourceChecker.java:206) org.apache.hadoop.hbase.ResourceCheckerJUnitListener.end(ResourceCheckerJUnitListener.java:165) org.apache.hadoop.hbase.ResourceCheckerJUnitListener.testFinished(ResourceCheckerJUnitListener.java:185) org.junit.runner.notification.SynchronizedRunListener.testFinished(SynchronizedRunListener.java:87) org.junit.runner.notification.RunNotifier$9.notifyListener(RunNotifier.java:225) org.junit.runner.notification.RunNotifier$SafeNotifier.run(RunNotifier.java:72) org.junit.runner.notification.RunNotifier.fireTestFinished(RunNotifier.java:222) org.junit.internal.runners.model.EachTestNotifier.fireTestFinished(EachTestNotifier.java:38) org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:372) org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) java.util.concurrent.FutureTask.run(FutureTask.java:266) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 3 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #56 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 11 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Container metrics unregistration java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: 788445821@qtp-1770429183-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:46353 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498) org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192) org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124) org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Potentially hanging thread: IPC Server handler 26 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 48 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: nioEventLoopGroup-10-1 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 11 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 25 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 31 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 43 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 40033 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-13 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Timer-45 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data6/current sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@411d3447 java.lang.Thread.sleep(Native Method) org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:3883) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-8 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 2 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #39 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 29 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 1214577836@qtp-669675177-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32801 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498) org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192) org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124) org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Potentially hanging thread: New I/O worker #65 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: ContainersLauncher #0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460) java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362) java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 15 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 3 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 14 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 3 on default port 40033 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: HFileArchiver-2 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: region-location-2 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1081) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 19 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 34777 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 5 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 38 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 6 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 27 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@582e6acc sun.nio.ch.ServerSocketChannelImpl.accept0(Native Method) sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:421) sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:249) sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:113) org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:145) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 16 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 45741 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: New I/O worker #67 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 17 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 15 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: org.apache.hadoop.util.JvmPauseMonitor$Monitor@70c378b1 java.lang.Thread.sleep(Native Method) org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:192) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #43 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 15 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 45 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-41 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 6 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 0 on default port 43257 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 45 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-57 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3828124c sun.nio.ch.ServerSocketChannelImpl.accept0(Native Method) sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:421) sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:249) sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:113) org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:145) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 46 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 1049017262@qtp-1974749666-0 java.lang.Object.wait(Native Method) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626) Potentially hanging thread: IPC Server handler 19 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 13 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 24 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #36 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 0 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: org.apache.hadoop.util.JvmPauseMonitor$Monitor@58ba879e java.lang.Thread.sleep(Native Method) org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:192) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-667153323_1 at /127.0.0.1:50242 [Waiting for operation #2] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read(BufferedInputStream.java:265) java.io.DataInputStream.readShort(DataInputStream.java:312) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:67) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:269) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 26 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 45831 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #46 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: pool-372-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Timer-42 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: ReadOnlyZKClient-127.0.0.1:52413@0x6fbf2c4a-EventThread sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:506) Potentially hanging thread: Timer-44 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: pool-395-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #44 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: pool-344-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 9 on default port 46621 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O server boss #68 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) org.jboss.netty.channel.socket.nio.NioServerBoss.select(NioServerBoss.java:163) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: org.apache.hadoop.util.JvmPauseMonitor$Monitor@4358a79b java.lang.Thread.sleep(Native Method) org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:192) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Thread[Thread-1363,5,FailOnTimeoutGroup] java.lang.Thread.sleep(Native Method) org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover.run(AbstractDelegationTokenSecretManager.java:694) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 16 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 0 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 12 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 20 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 5 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 17 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: DeletionService #2 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1081) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 5 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: pool-396-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #52 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: New I/O worker #66 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 10 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 40033 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 40215 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 28 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 29 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #64 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: org.apache.hadoop.util.JvmPauseMonitor$Monitor@5fa3c9eb java.lang.Thread.sleep(Native Method) org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:192) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-16 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 41 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 40215 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 43 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-58 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 6 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 20 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 0 on default port 45831 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 44 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 18 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 379127953@qtp-1240153792-0 java.lang.Object.wait(Native Method) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626) Potentially hanging thread: nioEventLoopGroup-14-1 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server idle connection scanner for port 36575 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 37 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-46 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 6 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 22 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 6 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 41523 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server idle connection scanner for port 45831 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: org.apache.hadoop.hdfs.server.blockmanagement.PendingReplicationBlocks$PendingReplicationMonitor@7ce9fc8f java.lang.Thread.sleep(Native Method) org.apache.hadoop.hdfs.server.blockmanagement.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:244) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: ReadOnlyZKClient-127.0.0.1:52413@0x6fbf2c4a-SendThread(127.0.0.1:52413) sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:345) org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1223) Potentially hanging thread: pool-400-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: ReadOnlyZKClient-127.0.0.1:52413@0x6fbf2c4a sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.DelayQueue.poll(DelayQueue.java:259) org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient.run(ReadOnlyZKClient.java:328) org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$493/2112819362.run(Unknown Source) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: DeletionService #3 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1081) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 9 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 25 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 1820115456@qtp-1816563711-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41911 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498) org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192) org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124) org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Potentially hanging thread: org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@61a12ee1 java.lang.Thread.sleep(Native Method) org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:533) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 2 on default port 41523 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #57 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Timer-60 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 0 on default port 40033 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 7 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #62 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 10 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Client (1886842325) connection to localhost/127.0.0.1:40033 from jenkins java.lang.Object.wait(Native Method) org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1053) org.apache.hadoop.ipc.Client$Connection.run(Client.java:1097) Potentially hanging thread: Timer-40 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@573270d5 java.lang.Thread.sleep(Native Method) org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:3841) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 0 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: New I/O worker #59 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.jboss.netty.channel.socket.nio.SelectorUtil.select(SelectorUtil.java:68) org.jboss.netty.channel.socket.nio.AbstractNioSelector.select(AbstractNioSelector.java:434) org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:212) org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89) org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108) org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 42 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 1 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_2129524288_16 at /127.0.0.1:36620 [Receiving block BP-2138422956-172.31.12.81-1691649115962:blk_1073741876_1052] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read1(BufferedInputStream.java:286) java.io.BufferedInputStream.read(BufferedInputStream.java:345) java.io.DataInputStream.read(DataInputStream.java:149) org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:209) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:211) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134) org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109) org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:528) org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:968) org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:877) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:166) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:103) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 30 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: pool-346-thread-1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 26 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server idle connection scanner for port 46627 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: CacheReplicationMonitor(118724632) sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2163) org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:181) Potentially hanging thread: ForkJoinPool-2-worker-0 sun.misc.Unsafe.park(Native Method) java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1824) java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1693) java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175) Potentially hanging thread: Timer-63 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 2 on default port 40215 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 4 on default port 35967 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 9 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 46 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-61 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 35 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-62 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 21 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-52 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 13 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 8 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: DataNode DiskChecker thread 1 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460) java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362) java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: LeaseRenewer:jenkins@localhost:40033 java.lang.Thread.sleep(Native Method) org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:412) org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$600(LeaseRenewer.java:76) org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:308) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: nioEventLoopGroup-12-1 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 3 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data2/current/BP-590518898-172.31.12.81-1691649173071 java.lang.Thread.sleep(Native Method) org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:205) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-667153323_1 at /127.0.0.1:40820 [Waiting for operation #2] sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.io.BufferedInputStream.fill(BufferedInputStream.java:246) java.io.BufferedInputStream.read(BufferedInputStream.java:265) java.io.DataInputStream.readShort(DataInputStream.java:312) org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:67) org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:269) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server idle connection scanner for port 46621 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/1bb09c49-cab0-8269-63a9-6a0bd264f12d/cluster_44464488-fb20-98cf-4aee-89b39555e1bc/dfs/data/data4/current/BP-590518898-172.31.12.81-1691649173071 java.lang.Thread.sleep(Native Method) org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:205) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: IPC Server handler 7 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 39 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 1975530141@qtp-800713920-0 java.lang.Object.wait(Native Method) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626) Potentially hanging thread: IPC Server handler 36 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: hconnection-0x21d01e4e-shared-pool-18 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: 335723368@qtp-539198283-0 java.lang.Object.wait(Native Method) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626) Potentially hanging thread: IPC Server handler 2 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 14 on default port 34777 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: BP-590518898-172.31.12.81-1691649173071 heartbeating to localhost/127.0.0.1:40033 java.lang.Object.wait(Native Method) org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:715) org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:851) java.lang.Thread.run(Thread.java:750) Potentially hanging thread: Timer-66 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 21 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 18 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 33 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 18 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 36 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 23 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 18 on default port 36575 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: Timer-43 java.lang.Object.wait(Native Method) java.util.TimerThread.mainLoop(Timer.java:552) java.util.TimerThread.run(Timer.java:505) Potentially hanging thread: IPC Server handler 20 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 42 on default port 45741 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 10 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 12 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 0 on default port 32987 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: IPC Server handler 13 on default port 46627 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: 577422739@qtp-800713920-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:37093 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498) org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192) org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124) org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708) org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Potentially hanging thread: IPC Server handler 13 on default port 34931 sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467) org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:294) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2820) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-10 sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.lang.Thread.run(Thread.java:750) - Thread LEAK? -, OpenFileDescriptor=1256 (was 903) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=60000 (was 60000), SystemLoadAverage=102 (was 91) - SystemLoadAverage LEAK? -, ProcessCount=174 (was 171) - ProcessCount LEAK? -, AvailableMemoryMB=922 (was 1984) 2023-08-10T06:33:21,116 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.ResourceChecker(130): Thread=1306 is superior to 500 2023-08-10T06:33:21,116 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.ResourceChecker(130): OpenFileDescriptor=1256 is superior to 1024 2023-08-10T06:33:21,124 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2023-08-10T06:33:21,124 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x1cbcc976 to 127.0.0.1:52413 2023-08-10T06:33:21,124 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:21,129 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2023-08-10T06:33:21,129 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] util.JVMClusterUtil(257): Found active master hash=1950382708, stopped=false 2023-08-10T06:33:21,130 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] master.ServerManager(910): Cluster shutdown requested of master=jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:33:21,131 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/2/running 2023-08-10T06:33:21,131 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/2/running 2023-08-10T06:33:21,131 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] procedure2.ProcedureExecutor(631): Stopping 2023-08-10T06:33:21,131 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:33:21,131 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:33:21,132 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:21,132 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] regionserver.HRegionServer(2513): ***** STOPPING region server 'jenkins-hbase3.apache.org,39491,1691649127997' ***** 2023-08-10T06:33:21,132 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/running 2023-08-10T06:33:21,132 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] regionserver.HRegionServer(2527): STOPPED: Shutdown requested 2023-08-10T06:33:21,132 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/running 2023-08-10T06:33:21,132 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1172): Stopping infoServer 2023-08-10T06:33:21,153 INFO [RS:0;jenkins-hbase3:39491 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.w.WebAppContext@2695d275{regionserver,/,null,STOPPED}{file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/regionserver} 2023-08-10T06:33:21,158 INFO [RS:0;jenkins-hbase3:39491 {}] server.AbstractConnector(383): Stopped ServerConnector@67bd5ab4{HTTP/1.1, (http/1.1)}{0.0.0.0:0} 2023-08-10T06:33:21,158 INFO [RS:0;jenkins-hbase3:39491 {}] session.HouseKeeper(149): node0 Stopped scavenging 2023-08-10T06:33:21,159 INFO [RS:0;jenkins-hbase3:39491 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.s.ServletContextHandler@487b472a{static,/static,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/static/,STOPPED} 2023-08-10T06:33:21,160 INFO [RS:0;jenkins-hbase3:39491 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.s.ServletContextHandler@eb06edd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop.log.dir/,STOPPED} 2023-08-10T06:33:21,163 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HeapMemoryManager(220): Stopping 2023-08-10T06:33:21,163 INFO [RS:0;jenkins-hbase3:39491 {}] flush.RegionServerFlushTableProcedureManager(117): Stopping region server flush procedure manager gracefully. 2023-08-10T06:33:21,163 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.LogRollRegionServerProcedureManager(106): Stopping RegionServerBackupManager gracefully. 2023-08-10T06:33:21,163 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2023-08-10T06:33:21,163 INFO [RS:0;jenkins-hbase3:39491 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2023-08-10T06:33:21,163 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(3524): Received CLOSE for 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:33:21,163 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1213): stopping server jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:33:21,164 DEBUG [RS:0;jenkins-hbase3:39491 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:21,165 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2023-08-10T06:33:21,165 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2023-08-10T06:33:21,165 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2023-08-10T06:33:21,165 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(3524): Received CLOSE for 1588230740 2023-08-10T06:33:21,166 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1668): Closing 86e6ed14dc5c14940dff214bf60e7e35, disabling compactions & flushes 2023-08-10T06:33:21,166 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1690): Closing region hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:33:21,166 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1711): Waiting without time limit for close lock on hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:33:21,166 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1778): Acquired close lock on hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. after waiting 0 ms 2023-08-10T06:33:21,166 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1579): Waiting on 2 regions to close 2023-08-10T06:33:21,166 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1788): Updates disabled for region hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:33:21,166 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1583): Online Regions={86e6ed14dc5c14940dff214bf60e7e35=hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35., 1588230740=hbase:meta,,1.1588230740} 2023-08-10T06:33:21,166 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2819): Flushing 86e6ed14dc5c14940dff214bf60e7e35 1/1 column families, dataSize=78 B heapSize=488 B 2023-08-10T06:33:21,167 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1668): Closing 1588230740, disabling compactions & flushes 2023-08-10T06:33:21,167 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1690): Closing region hbase:meta,,1.1588230740 2023-08-10T06:33:21,167 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1711): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2023-08-10T06:33:21,167 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1778): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2023-08-10T06:33:21,168 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1788): Updates disabled for region hbase:meta,,1.1588230740 2023-08-10T06:33:21,168 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2819): Flushing 1588230740 3/3 column families, dataSize=1.26 KB heapSize=2.89 KB 2023-08-10T06:33:21,171 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:33:21,175 INFO [regionserver/jenkins-hbase3:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2023-08-10T06:33:21,193 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35/.tmp/info/70a17b539e0047e59b3e422da07fd790 is 45, key is default/info:d/1691649130510/Put/seqid=0 2023-08-10T06:33:21,196 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/.tmp/info/a1729caac7b24556ae8c2f80cc277c59 is 143, key is hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35./info:regioninfo/1691649130387/Put/seqid=0 2023-08-10T06:33:21,200 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.17 KB at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/.tmp/info/a1729caac7b24556ae8c2f80cc277c59 2023-08-10T06:33:21,219 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: a1729caac7b24556ae8c2f80cc277c59 2023-08-10T06:33:21,230 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/.tmp/table/18d9fa78aebc41969708e61a810e408e is 51, key is hbase:namespace/table:state/1691649130398/Put/seqid=0 2023-08-10T06:33:21,371 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:33:21,420 INFO [regionserver/jenkins-hbase3:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2023-08-10T06:33:21,420 INFO [regionserver/jenkins-hbase3:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2023-08-10T06:33:21,572 DEBUG [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 86e6ed14dc5c14940dff214bf60e7e35 2023-08-10T06:33:21,598 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35/.tmp/info/70a17b539e0047e59b3e422da07fd790 2023-08-10T06:33:21,607 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 70a17b539e0047e59b3e422da07fd790 2023-08-10T06:33:21,608 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35/.tmp/info/70a17b539e0047e59b3e422da07fd790 as hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35/info/70a17b539e0047e59b3e422da07fd790 2023-08-10T06:33:21,613 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35/info/70a17b539e0047e59b3e422da07fd790, entries=2, sequenceid=6, filesize=4.9 K 2023-08-10T06:33:21,614 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3022): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 86e6ed14dc5c14940dff214bf60e7e35 in 448ms, sequenceid=6, compaction requested=false 2023-08-10T06:33:21,616 DEBUG [StoreFileCloser-86e6ed14dc5c14940dff214bf60e7e35-info-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 70a17b539e0047e59b3e422da07fd790 2023-08-10T06:33:21,621 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:44455/tmp/wal/data/hbase/namespace/86e6ed14dc5c14940dff214bf60e7e35/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2023-08-10T06:33:21,622 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:21,622 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1909): Closed hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:33:21,622 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1622): Region close journal for 86e6ed14dc5c14940dff214bf60e7e35: 2023-08-10T06:33:21,622 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1691649129590.86e6ed14dc5c14940dff214bf60e7e35. 2023-08-10T06:33:21,634 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=94 B at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/.tmp/table/18d9fa78aebc41969708e61a810e408e 2023-08-10T06:33:21,640 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 18d9fa78aebc41969708e61a810e408e 2023-08-10T06:33:21,640 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/.tmp/info/a1729caac7b24556ae8c2f80cc277c59 as hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/info/a1729caac7b24556ae8c2f80cc277c59 2023-08-10T06:33:21,645 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/info/a1729caac7b24556ae8c2f80cc277c59, entries=10, sequenceid=9, filesize=6.4 K 2023-08-10T06:33:21,646 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/.tmp/table/18d9fa78aebc41969708e61a810e408e as hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/table/18d9fa78aebc41969708e61a810e408e 2023-08-10T06:33:21,651 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/data/hbase/meta/1588230740/table/18d9fa78aebc41969708e61a810e408e, entries=2, sequenceid=9, filesize=5.1 K 2023-08-10T06:33:21,651 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3022): Finished flush of dataSize ~1.26 KB/1290, heapSize ~2.61 KB/2672, currentSize=0 B/0 for 1588230740 in 483ms, sequenceid=9, compaction requested=false 2023-08-10T06:33:21,654 DEBUG [StoreFileCloser-1588230740-info-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: a1729caac7b24556ae8c2f80cc277c59 2023-08-10T06:33:21,657 DEBUG [StoreFileCloser-1588230740-table-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 18d9fa78aebc41969708e61a810e408e 2023-08-10T06:33:21,662 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:44455/tmp/wal/data/hbase/meta/1588230740/recovered.edits/12.seqid, newMaxSeqId=12, maxSeqId=1 2023-08-10T06:33:21,663 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:21,663 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2023-08-10T06:33:21,663 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1909): Closed hbase:meta,,1.1588230740 2023-08-10T06:33:21,663 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1622): Region close journal for 1588230740: 2023-08-10T06:33:21,663 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2023-08-10T06:33:21,772 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1239): stopping server jenkins-hbase3.apache.org,39491,1691649127997; all regions closed. 2023-08-10T06:33:21,779 DEBUG [RS:0;jenkins-hbase3:39491 {}] wal.AbstractFSWAL(1062): Moved 1 WAL file(s) to /tmp/wal/oldWALs 2023-08-10T06:33:21,779 INFO [RS:0;jenkins-hbase3:39491 {}] wal.AbstractFSWAL(1065): Closed WAL: AsyncFSWAL jenkins-hbase3.apache.org%2C39491%2C1691649127997.meta:.meta(num 1691649129512) 2023-08-10T06:33:21,784 DEBUG [RS:0;jenkins-hbase3:39491 {}] wal.AbstractFSWAL(1062): Moved 1 WAL file(s) to /tmp/wal/oldWALs 2023-08-10T06:33:21,784 INFO [RS:0;jenkins-hbase3:39491 {}] wal.AbstractFSWAL(1065): Closed WAL: AsyncFSWAL jenkins-hbase3.apache.org%2C39491%2C1691649127997:(num 1691649128555) 2023-08-10T06:33:21,784 DEBUG [RS:0;jenkins-hbase3:39491 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:21,784 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.LeaseManager(133): Closed leases 2023-08-10T06:33:21,784 INFO [RS:0;jenkins-hbase3:39491 {}] hbase.ChoreService(370): Chore service for: regionserver/jenkins-hbase3:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2023-08-10T06:33:21,785 INFO [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2023-08-10T06:33:21,786 INFO [RS:0;jenkins-hbase3:39491 {}] ipc.NettyRpcServer(213): Stopping server on /172.31.12.81:39491 2023-08-10T06:33:21,792 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/2/rs/jenkins-hbase3.apache.org,39491,1691649127997 2023-08-10T06:33:21,792 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2/rs 2023-08-10T06:33:21,794 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [jenkins-hbase3.apache.org,39491,1691649127997] 2023-08-10T06:33:21,794 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing jenkins-hbase3.apache.org,39491,1691649127997; numProcessing=1 2023-08-10T06:33:21,796 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(224): Node /2/draining/jenkins-hbase3.apache.org,39491,1691649127997 already deleted, retry=false 2023-08-10T06:33:21,796 INFO [RegionServerTracker-0 {}] master.ServerManager(569): Cluster shutdown set; jenkins-hbase3.apache.org,39491,1691649127997 expired; onlineServers=0 2023-08-10T06:33:21,797 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2513): ***** STOPPING region server 'jenkins-hbase3.apache.org,38291,1691649127857' ***** 2023-08-10T06:33:21,797 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2527): STOPPED: Cluster shutdown set; onlineServer=0 2023-08-10T06:33:21,798 DEBUG [M:0;jenkins-hbase3:38291 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@a1b33d0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=jenkins-hbase3.apache.org/172.31.12.81:0 2023-08-10T06:33:21,798 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HRegionServer(1172): Stopping infoServer 2023-08-10T06:33:21,810 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/2/master 2023-08-10T06:33:21,810 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/2 2023-08-10T06:33:21,811 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Set watcher on znode that does not yet exist, /2/master 2023-08-10T06:33:21,892 INFO [M:0;jenkins-hbase3:38291 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.w.WebAppContext@20c7d6e9{master,/,null,STOPPED}{file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/master} 2023-08-10T06:33:21,893 INFO [M:0;jenkins-hbase3:38291 {}] server.AbstractConnector(383): Stopped ServerConnector@5443d879{HTTP/1.1, (http/1.1)}{0.0.0.0:0} 2023-08-10T06:33:21,893 INFO [M:0;jenkins-hbase3:38291 {}] session.HouseKeeper(149): node0 Stopped scavenging 2023-08-10T06:33:21,894 INFO [M:0;jenkins-hbase3:38291 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.s.ServletContextHandler@3f9b45fb{static,/static,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/static/,STOPPED} 2023-08-10T06:33:21,895 INFO [M:0;jenkins-hbase3:38291 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.s.ServletContextHandler@45ff7847{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/hadoop.log.dir/,STOPPED} 2023-08-10T06:33:21,895 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=None, state=Closed, path=null 2023-08-10T06:33:21,895 INFO [RS:0;jenkins-hbase3:39491 {}] regionserver.HRegionServer(1296): Exiting; stopping=jenkins-hbase3.apache.org,39491,1691649127997; zookeeper connection closed. 2023-08-10T06:33:21,895 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): regionserver:39491-0x10060ee705b0005, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=None, state=Closed, path=null 2023-08-10T06:33:21,896 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@77a8affb {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@77a8affb 2023-08-10T06:33:21,896 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HRegionServer(1213): stopping server jenkins-hbase3.apache.org,38291,1691649127857 2023-08-10T06:33:21,896 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HRegionServer(1239): stopping server jenkins-hbase3.apache.org,38291,1691649127857; all regions closed. 2023-08-10T06:33:21,896 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2023-08-10T06:33:21,896 DEBUG [M:0;jenkins-hbase3:38291 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:21,896 INFO [M:0;jenkins-hbase3:38291 {}] master.HMaster(1664): Stopping master jetty server 2023-08-10T06:33:21,897 INFO [M:0;jenkins-hbase3:38291 {}] server.AbstractConnector(383): Stopped ServerConnector@2e1f0112{HTTP/1.1, (http/1.1)}{0.0.0.0:0} 2023-08-10T06:33:21,897 DEBUG [M:0;jenkins-hbase3:38291 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2023-08-10T06:33:21,897 INFO [M:0;jenkins-hbase3:38291 {}] master.BackupLogCleaner(171): Stopping BackupLogCleaner 2023-08-10T06:33:21,898 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2023-08-10T06:33:21,898 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.large.0-1691649128322 {}] cleaner.HFileCleaner(306): Exit Thread[master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.large.0-1691649128322,5,FailOnTimeoutGroup] 2023-08-10T06:33:21,898 DEBUG [M:0;jenkins-hbase3:38291 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x72266fe4 to 127.0.0.1:52413 2023-08-10T06:33:21,898 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.small.0-1691649128322 {}] cleaner.HFileCleaner(306): Exit Thread[master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.small.0-1691649128322,5,FailOnTimeoutGroup] 2023-08-10T06:33:21,898 DEBUG [M:0;jenkins-hbase3:38291 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:21,898 DEBUG [M:0;jenkins-hbase3:38291 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2023-08-10T06:33:21,898 INFO [M:0;jenkins-hbase3:38291 {}] hbase.ChoreService(370): Chore service for: master/jenkins-hbase3:0 had [] on shutdown 2023-08-10T06:33:21,898 DEBUG [M:0;jenkins-hbase3:38291 {}] master.HMaster(1687): Stopping service threads 2023-08-10T06:33:21,898 INFO [M:0;jenkins-hbase3:38291 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2023-08-10T06:33:21,898 ERROR [M:0;jenkins-hbase3:38291 {}] procedure2.ProcedureExecutor(655): ThreadGroup java.lang.ThreadGroup[name=PEWorkerGroup,maxpri=10] contains running threads; null: See STDOUT java.lang.ThreadGroup[name=PEWorkerGroup,maxpri=10] 2023-08-10T06:33:21,899 INFO [M:0;jenkins-hbase3:38291 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2023-08-10T06:33:21,899 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2023-08-10T06:33:21,900 DEBUG [M:0;jenkins-hbase3:38291 {}] zookeeper.ZKUtil(347): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Unable to get data of znode /2/master because node does not exist (not an error) 2023-08-10T06:33:21,900 WARN [M:0;jenkins-hbase3:38291 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2023-08-10T06:33:21,900 INFO [M:0;jenkins-hbase3:38291 {}] assignment.AssignmentManager(382): Stopping assignment manager 2023-08-10T06:33:21,900 INFO [M:0;jenkins-hbase3:38291 {}] region.MasterRegion(178): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2023-08-10T06:33:21,901 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(1668): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2023-08-10T06:33:21,917 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(1690): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:33:21,917 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:33:21,917 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(1778): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2023-08-10T06:33:21,917 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(1788): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:33:21,917 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(2819): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=25.76 KB heapSize=33.30 KB 2023-08-10T06:33:21,932 DEBUG [M:0;jenkins-hbase3:38291 {}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/ec94eb3cf26f4e3aaaf6b6ea07de6820 is 88, key is hbase:meta,,1/info:sn/1691649129334/Put/seqid=0 2023-08-10T06:33:22,337 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=530 B at sequenceid=76 (bloomFilter=true), to=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/ec94eb3cf26f4e3aaaf6b6ea07de6820 2023-08-10T06:33:22,342 DEBUG [M:0;jenkins-hbase3:38291 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: ec94eb3cf26f4e3aaaf6b6ea07de6820 2023-08-10T06:33:22,355 DEBUG [M:0;jenkins-hbase3:38291 {}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a36de25d6ef740c6b6f8e84f1519131c is 241, key is \x00\x00\x00\x00\x00\x00\x00\x01/proc:d/1691649129569/Put/seqid=0 2023-08-10T06:33:22,760 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=25.17 KB at sequenceid=76 (bloomFilter=true), to=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a36de25d6ef740c6b6f8e84f1519131c 2023-08-10T06:33:22,765 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.StoreFileReader(536): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for a36de25d6ef740c6b6f8e84f1519131c 2023-08-10T06:33:22,765 DEBUG [M:0;jenkins-hbase3:38291 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: a36de25d6ef740c6b6f8e84f1519131c 2023-08-10T06:33:22,778 DEBUG [M:0;jenkins-hbase3:38291 {}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/382a6a2c21d64559981f5e889f686976 is 82, key is jenkins-hbase3.apache.org,39491,1691649127997/rs:state/1691649128389/Put/seqid=0 2023-08-10T06:33:23,182 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=76 (bloomFilter=true), to=hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/382a6a2c21d64559981f5e889f686976 2023-08-10T06:33:23,187 DEBUG [M:0;jenkins-hbase3:38291 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 382a6a2c21d64559981f5e889f686976 2023-08-10T06:33:23,187 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/ec94eb3cf26f4e3aaaf6b6ea07de6820 as hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/ec94eb3cf26f4e3aaaf6b6ea07de6820 2023-08-10T06:33:23,192 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/ec94eb3cf26f4e3aaaf6b6ea07de6820, entries=8, sequenceid=76, filesize=5.5 K 2023-08-10T06:33:23,193 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a36de25d6ef740c6b6f8e84f1519131c as hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/a36de25d6ef740c6b6f8e84f1519131c 2023-08-10T06:33:23,198 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.StoreFileReader(536): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for a36de25d6ef740c6b6f8e84f1519131c 2023-08-10T06:33:23,198 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/a36de25d6ef740c6b6f8e84f1519131c, entries=8, sequenceid=76, filesize=5.4 K 2023-08-10T06:33:23,199 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/382a6a2c21d64559981f5e889f686976 as hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/382a6a2c21d64559981f5e889f686976 2023-08-10T06:33:23,203 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:44455/user/jenkins/test-data/64f117f6-2dae-6f71-60f4-a6dd807940fa/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/382a6a2c21d64559981f5e889f686976, entries=1, sequenceid=76, filesize=5.1 K 2023-08-10T06:33:23,204 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(3022): Finished flush of dataSize ~25.76 KB/26381, heapSize ~33.01 KB/33800, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 1287ms, sequenceid=76, compaction requested=false 2023-08-10T06:33:23,206 DEBUG [StoreFileCloser-1595e783b53d99cd5eef43b6debb2682-info-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: ec94eb3cf26f4e3aaaf6b6ea07de6820 2023-08-10T06:33:23,209 DEBUG [StoreFileCloser-1595e783b53d99cd5eef43b6debb2682-proc-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: a36de25d6ef740c6b6f8e84f1519131c 2023-08-10T06:33:23,211 DEBUG [StoreFileCloser-1595e783b53d99cd5eef43b6debb2682-rs-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 382a6a2c21d64559981f5e889f686976 2023-08-10T06:33:23,213 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(1909): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:33:23,213 DEBUG [M:0;jenkins-hbase3:38291 {}] regionserver.HRegion(1622): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2023-08-10T06:33:23,216 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2023-08-10T06:33:23,216 INFO [M:0;jenkins-hbase3:38291 {}] flush.MasterFlushTableProcedureManager(81): stop: server shutting down. 2023-08-10T06:33:23,216 INFO [M:0;jenkins-hbase3:38291 {}] master.LogRollMasterProcedureManager(73): stop: server shutting down. 2023-08-10T06:33:23,217 INFO [M:0;jenkins-hbase3:38291 {}] ipc.NettyRpcServer(213): Stopping server on /172.31.12.81:38291 2023-08-10T06:33:23,219 DEBUG [M:0;jenkins-hbase3:38291 {}] zookeeper.RecoverableZooKeeper(224): Node /2/rs/jenkins-hbase3.apache.org,38291,1691649127857 already deleted, retry=false 2023-08-10T06:33:23,321 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=None, state=Closed, path=null 2023-08-10T06:33:23,321 INFO [M:0;jenkins-hbase3:38291 {}] regionserver.HRegionServer(1296): Exiting; stopping=jenkins-hbase3.apache.org,38291,1691649127857; zookeeper connection closed. 2023-08-10T06:33:23,321 DEBUG [Listener at localhost/36659-EventThread {}] zookeeper.ZKWatcher(604): master:38291-0x10060ee705b0004, quorum=127.0.0.1:52413, baseZNode=/2 Received ZooKeeper Event, type=None, state=Closed, path=null 2023-08-10T06:33:23,324 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] datanode.DirectoryScanner(537): DirectoryScanner: shutdown has been called 2023-08-10T06:33:23,333 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0 2023-08-10T06:33:23,438 WARN [BP-564232180-172.31.12.81-1691649127005 heartbeating to localhost/127.0.0.1:44455 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2023-08-10T06:33:23,438 WARN [BP-564232180-172.31.12.81-1691649127005 heartbeating to localhost/127.0.0.1:44455 {}] datanode.BPServiceActor(862): Ending block pool service for: Block pool BP-564232180-172.31.12.81-1691649127005 (Datanode Uuid b86086af-0069-40ac-b54c-a5ce0bf8379b) service to localhost/127.0.0.1:44455 2023-08-10T06:33:23,440 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/cluster_8430befc-bdf7-6f28-aaa6-942b3e94ea59/dfs/data/data5/current/BP-564232180-172.31.12.81-1691649127005 {}] fs.CachingGetSpaceUsed$RefreshThread(211): Thread Interrupted waiting to refresh disk information: sleep interrupted 2023-08-10T06:33:23,440 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/cluster_8430befc-bdf7-6f28-aaa6-942b3e94ea59/dfs/data/data6/current/BP-564232180-172.31.12.81-1691649127005 {}] fs.CachingGetSpaceUsed$RefreshThread(211): Thread Interrupted waiting to refresh disk information: sleep interrupted 2023-08-10T06:33:23,444 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] datanode.DirectoryScanner(537): DirectoryScanner: shutdown has been called 2023-08-10T06:33:23,452 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0 2023-08-10T06:33:23,558 WARN [BP-564232180-172.31.12.81-1691649127005 heartbeating to localhost/127.0.0.1:44455 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2023-08-10T06:33:23,558 WARN [BP-564232180-172.31.12.81-1691649127005 heartbeating to localhost/127.0.0.1:44455 {}] datanode.BPServiceActor(862): Ending block pool service for: Block pool BP-564232180-172.31.12.81-1691649127005 (Datanode Uuid 868b385d-1a59-4611-8f2d-3e9d2b8e4e1e) service to localhost/127.0.0.1:44455 2023-08-10T06:33:23,559 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/cluster_8430befc-bdf7-6f28-aaa6-942b3e94ea59/dfs/data/data3/current/BP-564232180-172.31.12.81-1691649127005 {}] fs.CachingGetSpaceUsed$RefreshThread(211): Thread Interrupted waiting to refresh disk information: sleep interrupted 2023-08-10T06:33:23,559 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/cluster_8430befc-bdf7-6f28-aaa6-942b3e94ea59/dfs/data/data4/current/BP-564232180-172.31.12.81-1691649127005 {}] fs.CachingGetSpaceUsed$RefreshThread(211): Thread Interrupted waiting to refresh disk information: sleep interrupted 2023-08-10T06:33:23,563 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] datanode.DirectoryScanner(537): DirectoryScanner: shutdown has been called 2023-08-10T06:33:23,571 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0 2023-08-10T06:33:23,676 WARN [BP-564232180-172.31.12.81-1691649127005 heartbeating to localhost/127.0.0.1:44455 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2023-08-10T06:33:23,676 WARN [BP-564232180-172.31.12.81-1691649127005 heartbeating to localhost/127.0.0.1:44455 {}] datanode.BPServiceActor(862): Ending block pool service for: Block pool BP-564232180-172.31.12.81-1691649127005 (Datanode Uuid 294aae14-1491-4c05-bc43-e02619891c37) service to localhost/127.0.0.1:44455 2023-08-10T06:33:23,677 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/cluster_8430befc-bdf7-6f28-aaa6-942b3e94ea59/dfs/data/data1/current/BP-564232180-172.31.12.81-1691649127005 {}] fs.CachingGetSpaceUsed$RefreshThread(211): Thread Interrupted waiting to refresh disk information: sleep interrupted 2023-08-10T06:33:23,677 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/98b850ed-3b99-13d8-2ab2-9db2775856ec/cluster_8430befc-bdf7-6f28-aaa6-942b3e94ea59/dfs/data/data2/current/BP-564232180-172.31.12.81-1691649127005 {}] fs.CachingGetSpaceUsed$RefreshThread(211): Thread Interrupted waiting to refresh disk information: sleep interrupted 2023-08-10T06:33:23,716 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0 2023-08-10T06:33:23,856 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2023-08-10T06:33:23,856 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2023-08-10T06:33:23,856 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:33:23,856 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x2e3dc5ff to 127.0.0.1:52413 2023-08-10T06:33:23,856 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:23,857 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2023-08-10T06:33:23,857 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] util.JVMClusterUtil(257): Found active master hash=1230679401, stopped=false 2023-08-10T06:33:23,857 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] master.ServerManager(910): Cluster shutdown requested of master=jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:33:23,858 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/running 2023-08-10T06:33:23,858 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/running 2023-08-10T06:33:23,858 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:33:23,858 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:33:23,858 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] procedure2.ProcedureExecutor(631): Stopping 2023-08-10T06:33:23,859 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/running 2023-08-10T06:33:23,859 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/running 2023-08-10T06:33:23,859 DEBUG [Listener at jenkins-hbase3.apache.org/41523 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:23,859 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] regionserver.HRegionServer(2513): ***** STOPPING region server 'jenkins-hbase3.apache.org,34991,1691649122119' ***** 2023-08-10T06:33:23,859 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] regionserver.HRegionServer(2527): STOPPED: Shutdown requested 2023-08-10T06:33:23,859 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1172): Stopping infoServer 2023-08-10T06:33:23,864 INFO [RS:0;jenkins-hbase3:34991 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.w.WebAppContext@4c765c3c{regionserver,/,null,STOPPED}{file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/regionserver} 2023-08-10T06:33:23,865 INFO [RS:0;jenkins-hbase3:34991 {}] server.AbstractConnector(383): Stopped ServerConnector@5233fa39{HTTP/1.1, (http/1.1)}{0.0.0.0:0} 2023-08-10T06:33:23,865 INFO [RS:0;jenkins-hbase3:34991 {}] session.HouseKeeper(149): node0 Stopped scavenging 2023-08-10T06:33:23,866 INFO [RS:0;jenkins-hbase3:34991 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.s.ServletContextHandler@2b2d8bca{static,/static,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/static/,STOPPED} 2023-08-10T06:33:23,867 INFO [RS:0;jenkins-hbase3:34991 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.s.ServletContextHandler@3c8b8bc0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.log.dir/,STOPPED} 2023-08-10T06:33:23,868 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HeapMemoryManager(220): Stopping 2023-08-10T06:33:23,868 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2023-08-10T06:33:23,868 INFO [RS:0;jenkins-hbase3:34991 {}] flush.RegionServerFlushTableProcedureManager(117): Stopping region server flush procedure manager gracefully. 2023-08-10T06:33:23,868 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.LogRollRegionServerProcedureManager(106): Stopping RegionServerBackupManager gracefully. 2023-08-10T06:33:23,868 INFO [RS:0;jenkins-hbase3:34991 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2023-08-10T06:33:23,869 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3524): Received CLOSE for c11396d9b009ba78cea809571c6e38e7 2023-08-10T06:33:23,869 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3524): Received CLOSE for 7ecca0e4e0228e4c7a687e00712cb8f1 2023-08-10T06:33:23,869 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3524): Received CLOSE for 66486f64c0e50e4c038f3174e657e6c0 2023-08-10T06:33:23,870 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1668): Closing c11396d9b009ba78cea809571c6e38e7, disabling compactions & flushes 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3524): Received CLOSE for de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:33:23,870 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1690): Closing region ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3524): Received CLOSE for 7f5405ff46e3916fa2d5dcf037138bc3 2023-08-10T06:33:23,870 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1711): Waiting without time limit for close lock on ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3524): Received CLOSE for 5a53179ddca6d17f3904c3231abeb782 2023-08-10T06:33:23,870 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1778): Acquired close lock on ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. after waiting 0 ms 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3524): Received CLOSE for 333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:33:23,870 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1788): Updates disabled for region ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1213): stopping server jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] client.ConnectionImplementation(2114): Closing master protocol: MasterService 2023-08-10T06:33:23,870 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2819): Flushing c11396d9b009ba78cea809571c6e38e7 1/1 column families, dataSize=3.17 KB heapSize=11 KB 2023-08-10T06:33:23,870 DEBUG [RS:0;jenkins-hbase3:34991 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2023-08-10T06:33:23,870 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(3524): Received CLOSE for 1588230740 2023-08-10T06:33:23,871 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1579): Waiting on 8 regions to close 2023-08-10T06:33:23,871 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1583): Online Regions={c11396d9b009ba78cea809571c6e38e7=ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7., 7ecca0e4e0228e4c7a687e00712cb8f1=ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1., 66486f64c0e50e4c038f3174e657e6c0=ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0., 1588230740=hbase:meta,,1.1588230740, de2170c3441572f6a0fca17a49c93cc1=backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1., 7f5405ff46e3916fa2d5dcf037138bc3=test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3., 5a53179ddca6d17f3904c3231abeb782=backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782., 333dde743887b960947ea82e4c5cfa5e=hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e.} 2023-08-10T06:33:23,872 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1668): Closing 1588230740, disabling compactions & flushes 2023-08-10T06:33:23,872 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 333dde743887b960947ea82e4c5cfa5e, 5a53179ddca6d17f3904c3231abeb782, 66486f64c0e50e4c038f3174e657e6c0, 7ecca0e4e0228e4c7a687e00712cb8f1, 7f5405ff46e3916fa2d5dcf037138bc3, c11396d9b009ba78cea809571c6e38e7, de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:33:23,872 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1690): Closing region hbase:meta,,1.1588230740 2023-08-10T06:33:23,872 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1711): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2023-08-10T06:33:23,873 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1778): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2023-08-10T06:33:23,873 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1788): Updates disabled for region hbase:meta,,1.1588230740 2023-08-10T06:33:23,873 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2819): Flushing 1588230740 3/3 column families, dataSize=13.51 KB heapSize=24 KB 2023-08-10T06:33:23,877 INFO [regionserver/jenkins-hbase3:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2023-08-10T06:33:23,892 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7/.tmp/f/cdb9563d04614878a5415d3e5facff8a is 37, key is row10/f:q1/1691649138984/Put/seqid=0 2023-08-10T06:33:23,897 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=3.17 KB at sequenceid=103 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7/.tmp/f/cdb9563d04614878a5415d3e5facff8a 2023-08-10T06:33:23,897 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/info/eb25f121ea7c4eef8792d5a2785eb0f1 is 159, key is ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7./info:regioninfo/1691649138669/Put/seqid=0 2023-08-10T06:33:23,904 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: cdb9563d04614878a5415d3e5facff8a 2023-08-10T06:33:23,905 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7/.tmp/f/cdb9563d04614878a5415d3e5facff8a as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7/f/cdb9563d04614878a5415d3e5facff8a 2023-08-10T06:33:23,910 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7/f/cdb9563d04614878a5415d3e5facff8a, entries=99, sequenceid=103, filesize=8.2 K 2023-08-10T06:33:23,911 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3022): Finished flush of dataSize ~3.17 KB/3247, heapSize ~10.98 KB/11248, currentSize=0 B/0 for c11396d9b009ba78cea809571c6e38e7 in 41ms, sequenceid=103, compaction requested=false 2023-08-10T06:33:23,914 DEBUG [StoreFileCloser-c11396d9b009ba78cea809571c6e38e7-f-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: cdb9563d04614878a5415d3e5facff8a 2023-08-10T06:33:23,919 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns2/test-16916491351021/c11396d9b009ba78cea809571c6e38e7/recovered.edits/106.seqid, newMaxSeqId=106, maxSeqId=1 2023-08-10T06:33:23,919 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:23,919 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1909): Closed ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:33:23,920 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1622): Region close journal for c11396d9b009ba78cea809571c6e38e7: 2023-08-10T06:33:23,920 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed ns2:test-16916491351021,,1691649137855.c11396d9b009ba78cea809571c6e38e7. 2023-08-10T06:33:23,921 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1668): Closing 7ecca0e4e0228e4c7a687e00712cb8f1, disabling compactions & flushes 2023-08-10T06:33:23,921 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1690): Closing region ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:33:23,921 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1711): Waiting without time limit for close lock on ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:33:23,921 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1778): Acquired close lock on ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. after waiting 0 ms 2023-08-10T06:33:23,921 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1788): Updates disabled for region ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:33:23,924 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns3/test-16916491351022/7ecca0e4e0228e4c7a687e00712cb8f1/recovered.edits/4.seqid, newMaxSeqId=4, maxSeqId=1 2023-08-10T06:33:23,924 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:23,924 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1909): Closed ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:33:23,924 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1622): Region close journal for 7ecca0e4e0228e4c7a687e00712cb8f1: 2023-08-10T06:33:23,924 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed ns3:test-16916491351022,,1691649139060.7ecca0e4e0228e4c7a687e00712cb8f1. 2023-08-10T06:33:23,925 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1668): Closing 66486f64c0e50e4c038f3174e657e6c0, disabling compactions & flushes 2023-08-10T06:33:23,925 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1690): Closing region ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:33:23,925 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1711): Waiting without time limit for close lock on ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:33:23,925 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1778): Acquired close lock on ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. after waiting 0 ms 2023-08-10T06:33:23,925 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1788): Updates disabled for region ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:33:23,928 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/ns4/test-16916491351023/66486f64c0e50e4c038f3174e657e6c0/recovered.edits/4.seqid, newMaxSeqId=4, maxSeqId=1 2023-08-10T06:33:23,928 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:23,929 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1909): Closed ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:33:23,929 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1622): Region close journal for 66486f64c0e50e4c038f3174e657e6c0: 2023-08-10T06:33:23,929 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed ns4:test-16916491351023,,1691649141183.66486f64c0e50e4c038f3174e657e6c0. 2023-08-10T06:33:23,929 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1668): Closing de2170c3441572f6a0fca17a49c93cc1, disabling compactions & flushes 2023-08-10T06:33:23,929 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1690): Closing region backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:33:23,929 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1711): Waiting without time limit for close lock on backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:33:23,929 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1778): Acquired close lock on backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. after waiting 0 ms 2023-08-10T06:33:23,929 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1788): Updates disabled for region backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:33:23,929 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2819): Flushing de2170c3441572f6a0fca17a49c93cc1 2/2 column families, dataSize=985 B heapSize=1.97 KB 2023-08-10T06:33:23,932 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/meta/74ae7e2c6d7144a2a1942a45b2a1b8c8 is 169, key is trslm:hdfs://localhost:44455/backupUT\x00test-1691649135102/meta:log-roll-map/1691649172613/Put/seqid=0 2023-08-10T06:33:23,988 INFO [regionserver/jenkins-hbase3:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2023-08-10T06:33:23,988 INFO [regionserver/jenkins-hbase3:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2023-08-10T06:33:24,073 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 333dde743887b960947ea82e4c5cfa5e, 5a53179ddca6d17f3904c3231abeb782, 7f5405ff46e3916fa2d5dcf037138bc3, de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:33:24,273 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 333dde743887b960947ea82e4c5cfa5e, 5a53179ddca6d17f3904c3231abeb782, 7f5405ff46e3916fa2d5dcf037138bc3, de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:33:24,303 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=12.12 KB at sequenceid=63 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/info/eb25f121ea7c4eef8792d5a2785eb0f1 2023-08-10T06:33:24,309 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: eb25f121ea7c4eef8792d5a2785eb0f1 2023-08-10T06:33:24,327 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/rep_barrier/afdf01170ad04a3ca4919e92a073538e is 90, key is table1,,1691649157953.95cd117c0618eb0283c218ddc0657261./rep_barrier:/1691649163343/DeleteFamily/seqid=0 2023-08-10T06:33:24,336 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=451 B at sequenceid=27 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/meta/74ae7e2c6d7144a2a1942a45b2a1b8c8 2023-08-10T06:33:24,341 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 74ae7e2c6d7144a2a1942a45b2a1b8c8 2023-08-10T06:33:24,344 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/5688996e6d2a4dce8a24f567a0239555 is 310, key is session:backup_1691649163473/session:context/1691649173035/Put/seqid=0 2023-08-10T06:33:24,473 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 333dde743887b960947ea82e4c5cfa5e, 5a53179ddca6d17f3904c3231abeb782, 7f5405ff46e3916fa2d5dcf037138bc3, de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:33:24,673 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 333dde743887b960947ea82e4c5cfa5e, 5a53179ddca6d17f3904c3231abeb782, 7f5405ff46e3916fa2d5dcf037138bc3, de2170c3441572f6a0fca17a49c93cc1 2023-08-10T06:33:24,734 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=172 B at sequenceid=63 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/rep_barrier/afdf01170ad04a3ca4919e92a073538e 2023-08-10T06:33:24,735 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2001): Auth successful for appattempt_1691649176425_0001_000001 (auth:SIMPLE) 2023-08-10T06:33:24,747 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: afdf01170ad04a3ca4919e92a073538e 2023-08-10T06:33:24,748 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=534 B at sequenceid=27 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/5688996e6d2a4dce8a24f567a0239555 2023-08-10T06:33:24,755 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 5688996e6d2a4dce8a24f567a0239555 2023-08-10T06:33:24,756 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/meta/74ae7e2c6d7144a2a1942a45b2a1b8c8 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/meta/74ae7e2c6d7144a2a1942a45b2a1b8c8 2023-08-10T06:33:24,762 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/meta/74ae7e2c6d7144a2a1942a45b2a1b8c8, entries=4, sequenceid=27, filesize=5.5 K 2023-08-10T06:33:24,763 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/.tmp/session/5688996e6d2a4dce8a24f567a0239555 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/5688996e6d2a4dce8a24f567a0239555 2023-08-10T06:33:24,767 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/table/3dea815411a741568b98fc47b1af0e2a is 84, key is table1,,1691649157953.95cd117c0618eb0283c218ddc0657261./table:/1691649163343/DeleteFamily/seqid=0 2023-08-10T06:33:24,771 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/session/5688996e6d2a4dce8a24f567a0239555, entries=2, sequenceid=27, filesize=5.3 K 2023-08-10T06:33:24,772 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3022): Finished flush of dataSize ~985 B/985, heapSize ~1.94 KB/1984, currentSize=0 B/0 for de2170c3441572f6a0fca17a49c93cc1 in 843ms, sequenceid=27, compaction requested=true 2023-08-10T06:33:24,774 DEBUG [StoreFileCloser-de2170c3441572f6a0fca17a49c93cc1-meta-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 53df5bd3efe64cb28d0a568f4284b10b 2023-08-10T06:33:24,777 DEBUG [StoreFileCloser-de2170c3441572f6a0fca17a49c93cc1-meta-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 74ae7e2c6d7144a2a1942a45b2a1b8c8 2023-08-10T06:33:24,781 DEBUG [StoreFileCloser-de2170c3441572f6a0fca17a49c93cc1-session-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: a7aa6b78efc846bebb42d595b96dfe44 2023-08-10T06:33:24,785 DEBUG [StoreFileCloser-de2170c3441572f6a0fca17a49c93cc1-session-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 397d772dc94c4e45be47d9f2f04a3258 2023-08-10T06:33:24,788 DEBUG [StoreFileCloser-de2170c3441572f6a0fca17a49c93cc1-session-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 5688996e6d2a4dce8a24f567a0239555 2023-08-10T06:33:24,797 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system/de2170c3441572f6a0fca17a49c93cc1/recovered.edits/30.seqid, newMaxSeqId=30, maxSeqId=1 2023-08-10T06:33:24,797 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:24,798 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1909): Closed backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:33:24,798 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1622): Region close journal for de2170c3441572f6a0fca17a49c93cc1: 2023-08-10T06:33:24,798 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed backup:system,,1691649143550.de2170c3441572f6a0fca17a49c93cc1. 2023-08-10T06:33:24,798 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1668): Closing 7f5405ff46e3916fa2d5dcf037138bc3, disabling compactions & flushes 2023-08-10T06:33:24,798 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1690): Closing region test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:33:24,798 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1711): Waiting without time limit for close lock on test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:33:24,798 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1778): Acquired close lock on test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. after waiting 0 ms 2023-08-10T06:33:24,798 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1788): Updates disabled for region test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:33:24,799 DEBUG [StoreFileCloser-7f5405ff46e3916fa2d5dcf037138bc3-f-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 317474dc8ece4aa0a1b9b2a97a3c57de 2023-08-10T06:33:24,804 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/default/test-1691649135102/7f5405ff46e3916fa2d5dcf037138bc3/recovered.edits/107.seqid, newMaxSeqId=107, maxSeqId=1 2023-08-10T06:33:24,805 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:24,805 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1909): Closed test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:33:24,805 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1622): Region close journal for 7f5405ff46e3916fa2d5dcf037138bc3: 2023-08-10T06:33:24,805 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed test-1691649135102,,1691649135617.7f5405ff46e3916fa2d5dcf037138bc3. 2023-08-10T06:33:24,806 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1668): Closing 5a53179ddca6d17f3904c3231abeb782, disabling compactions & flushes 2023-08-10T06:33:24,806 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1690): Closing region backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:33:24,806 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1711): Waiting without time limit for close lock on backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:33:24,806 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1778): Acquired close lock on backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. after waiting 0 ms 2023-08-10T06:33:24,806 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1788): Updates disabled for region backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:33:24,809 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/backup/system_bulk/5a53179ddca6d17f3904c3231abeb782/recovered.edits/4.seqid, newMaxSeqId=4, maxSeqId=1 2023-08-10T06:33:24,809 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:24,810 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1909): Closed backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:33:24,810 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1622): Region close journal for 5a53179ddca6d17f3904c3231abeb782: 2023-08-10T06:33:24,810 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed backup:system_bulk,,1691649145685.5a53179ddca6d17f3904c3231abeb782. 2023-08-10T06:33:24,811 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1668): Closing 333dde743887b960947ea82e4c5cfa5e, disabling compactions & flushes 2023-08-10T06:33:24,811 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1690): Closing region hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:33:24,811 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1711): Waiting without time limit for close lock on hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:33:24,811 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1778): Acquired close lock on hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. after waiting 0 ms 2023-08-10T06:33:24,811 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1788): Updates disabled for region hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:33:24,811 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2819): Flushing 333dde743887b960947ea82e4c5cfa5e 1/1 column families, dataSize=249 B heapSize=1.02 KB 2023-08-10T06:33:24,824 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e/.tmp/info/3fbd6fad373f442ebde8b0fd2e5ab854 is 45, key is default/info:d/1691649126849/Put/seqid=0 2023-08-10T06:33:24,874 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1579): Waiting on 2 regions to close 2023-08-10T06:33:24,874 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1583): Online Regions={1588230740=hbase:meta,,1.1588230740, 333dde743887b960947ea82e4c5cfa5e=hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e.} 2023-08-10T06:33:24,874 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:33:25,074 DEBUG [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1609): Waiting on 1588230740, 333dde743887b960947ea82e4c5cfa5e 2023-08-10T06:33:25,184 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.22 KB at sequenceid=63 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/table/3dea815411a741568b98fc47b1af0e2a 2023-08-10T06:33:25,190 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 3dea815411a741568b98fc47b1af0e2a 2023-08-10T06:33:25,191 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/info/eb25f121ea7c4eef8792d5a2785eb0f1 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/info/eb25f121ea7c4eef8792d5a2785eb0f1 2023-08-10T06:33:25,195 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/info/eb25f121ea7c4eef8792d5a2785eb0f1, entries=74, sequenceid=63, filesize=14.4 K 2023-08-10T06:33:25,196 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/rep_barrier/afdf01170ad04a3ca4919e92a073538e as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/rep_barrier/afdf01170ad04a3ca4919e92a073538e 2023-08-10T06:33:25,201 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/rep_barrier/afdf01170ad04a3ca4919e92a073538e, entries=2, sequenceid=63, filesize=5.4 K 2023-08-10T06:33:25,201 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/.tmp/table/3dea815411a741568b98fc47b1af0e2a as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/table/3dea815411a741568b98fc47b1af0e2a 2023-08-10T06:33:25,206 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/table/3dea815411a741568b98fc47b1af0e2a, entries=17, sequenceid=63, filesize=6.0 K 2023-08-10T06:33:25,207 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3022): Finished flush of dataSize ~13.51 KB/13834, heapSize ~23.95 KB/24528, currentSize=0 B/0 for 1588230740 in 1334ms, sequenceid=63, compaction requested=false 2023-08-10T06:33:25,210 DEBUG [StoreFileCloser-1588230740-info-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: eb25f121ea7c4eef8792d5a2785eb0f1 2023-08-10T06:33:25,213 DEBUG [StoreFileCloser-1588230740-rep_barrier-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: afdf01170ad04a3ca4919e92a073538e 2023-08-10T06:33:25,216 DEBUG [StoreFileCloser-1588230740-table-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 3dea815411a741568b98fc47b1af0e2a 2023-08-10T06:33:25,221 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/meta/1588230740/recovered.edits/66.seqid, newMaxSeqId=66, maxSeqId=1 2023-08-10T06:33:25,221 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:25,222 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2023-08-10T06:33:25,222 INFO [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1909): Closed hbase:meta,,1.1588230740 2023-08-10T06:33:25,222 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1622): Region close journal for 1588230740: 2023-08-10T06:33:25,222 DEBUG [RS_CLOSE_META-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2023-08-10T06:33:25,228 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=249 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e/.tmp/info/3fbd6fad373f442ebde8b0fd2e5ab854 2023-08-10T06:33:25,232 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 3fbd6fad373f442ebde8b0fd2e5ab854 2023-08-10T06:33:25,233 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e/.tmp/info/3fbd6fad373f442ebde8b0fd2e5ab854 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e/info/3fbd6fad373f442ebde8b0fd2e5ab854 2023-08-10T06:33:25,237 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e/info/3fbd6fad373f442ebde8b0fd2e5ab854, entries=7, sequenceid=11, filesize=5.1 K 2023-08-10T06:33:25,238 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3022): Finished flush of dataSize ~249 B/249, heapSize ~1.01 KB/1032, currentSize=0 B/0 for 333dde743887b960947ea82e4c5cfa5e in 427ms, sequenceid=11, compaction requested=false 2023-08-10T06:33:25,240 DEBUG [StoreFileCloser-333dde743887b960947ea82e4c5cfa5e-info-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 3fbd6fad373f442ebde8b0fd2e5ab854 2023-08-10T06:33:25,247 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/data/hbase/namespace/333dde743887b960947ea82e4c5cfa5e/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2023-08-10T06:33:25,248 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.backup.BackupObserver 2023-08-10T06:33:25,248 INFO [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1909): Closed hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:33:25,248 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1622): Region close journal for 333dde743887b960947ea82e4c5cfa5e: 2023-08-10T06:33:25,248 DEBUG [RS_CLOSE_REGION-regionserver/jenkins-hbase3:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1691649125464.333dde743887b960947ea82e4c5cfa5e. 2023-08-10T06:33:25,274 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1239): stopping server jenkins-hbase3.apache.org,34991,1691649122119; all regions closed. 2023-08-10T06:33:25,280 DEBUG [RS:0;jenkins-hbase3:34991 {}] wal.AbstractFSWAL(1062): Moved 3 WAL file(s) to /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/oldWALs 2023-08-10T06:33:25,280 INFO [RS:0;jenkins-hbase3:34991 {}] wal.AbstractFSWAL(1065): Closed WAL: AsyncFSWAL jenkins-hbase3.apache.org%2C34991%2C1691649122119.meta:.meta(num 1691649167659) 2023-08-10T06:33:25,286 DEBUG [RS:0;jenkins-hbase3:34991 {}] wal.AbstractFSWAL(1062): Moved 3 WAL file(s) to /user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/oldWALs 2023-08-10T06:33:25,286 INFO [RS:0;jenkins-hbase3:34991 {}] wal.AbstractFSWAL(1065): Closed WAL: AsyncFSWAL jenkins-hbase3.apache.org%2C34991%2C1691649122119:(num 1691649167640) 2023-08-10T06:33:25,286 DEBUG [RS:0;jenkins-hbase3:34991 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:25,286 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.LeaseManager(133): Closed leases 2023-08-10T06:33:25,287 INFO [RS:0;jenkins-hbase3:34991 {}] hbase.ChoreService(370): Chore service for: regionserver/jenkins-hbase3:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2023-08-10T06:33:25,287 INFO [regionserver/jenkins-hbase3:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2023-08-10T06:33:25,288 INFO [RS:0;jenkins-hbase3:34991 {}] ipc.NettyRpcServer(213): Stopping server on /172.31.12.81:34991 2023-08-10T06:33:25,293 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/rs/jenkins-hbase3.apache.org,34991,1691649122119 2023-08-10T06:33:25,293 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1/rs 2023-08-10T06:33:25,300 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [jenkins-hbase3.apache.org,34991,1691649122119] 2023-08-10T06:33:25,300 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing jenkins-hbase3.apache.org,34991,1691649122119; numProcessing=1 2023-08-10T06:33:25,301 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(224): Node /1/draining/jenkins-hbase3.apache.org,34991,1691649122119 already deleted, retry=false 2023-08-10T06:33:25,301 INFO [RegionServerTracker-0 {}] master.ServerManager(569): Cluster shutdown set; jenkins-hbase3.apache.org,34991,1691649122119 expired; onlineServers=0 2023-08-10T06:33:25,301 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2513): ***** STOPPING region server 'jenkins-hbase3.apache.org,42279,1691649119143' ***** 2023-08-10T06:33:25,301 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2527): STOPPED: Cluster shutdown set; onlineServer=0 2023-08-10T06:33:25,302 DEBUG [M:0;jenkins-hbase3:42279 {}] ipc.AbstractRpcClient(197): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@52a96e8a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=jenkins-hbase3.apache.org/172.31.12.81:0 2023-08-10T06:33:25,302 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HRegionServer(1172): Stopping infoServer 2023-08-10T06:33:25,318 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/1/master 2023-08-10T06:33:25,318 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/1 2023-08-10T06:33:25,318 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Set watcher on znode that does not yet exist, /1/master 2023-08-10T06:33:25,337 INFO [M:0;jenkins-hbase3:42279 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.w.WebAppContext@6bc05b4f{master,/,null,STOPPED}{file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/master} 2023-08-10T06:33:25,337 INFO [M:0;jenkins-hbase3:42279 {}] server.AbstractConnector(383): Stopped ServerConnector@46005c78{HTTP/1.1, (http/1.1)}{0.0.0.0:0} 2023-08-10T06:33:25,337 INFO [M:0;jenkins-hbase3:42279 {}] session.HouseKeeper(149): node0 Stopped scavenging 2023-08-10T06:33:25,339 INFO [M:0;jenkins-hbase3:42279 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.s.ServletContextHandler@67965e92{static,/static,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/classes/hbase-webapps/static/,STOPPED} 2023-08-10T06:33:25,340 INFO [M:0;jenkins-hbase3:42279 {}] handler.ContextHandler(1159): Stopped o.a.h.t.o.e.j.s.ServletContextHandler@4f27c993{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/hadoop.log.dir/,STOPPED} 2023-08-10T06:33:25,340 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HRegionServer(1213): stopping server jenkins-hbase3.apache.org,42279,1691649119143 2023-08-10T06:33:25,340 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HRegionServer(1239): stopping server jenkins-hbase3.apache.org,42279,1691649119143; all regions closed. 2023-08-10T06:33:25,340 DEBUG [M:0;jenkins-hbase3:42279 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:25,340 INFO [M:0;jenkins-hbase3:42279 {}] master.HMaster(1664): Stopping master jetty server 2023-08-10T06:33:25,341 INFO [M:0;jenkins-hbase3:42279 {}] server.AbstractConnector(383): Stopped ServerConnector@696643d9{HTTP/1.1, (http/1.1)}{0.0.0.0:0} 2023-08-10T06:33:25,341 DEBUG [M:0;jenkins-hbase3:42279 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2023-08-10T06:33:25,341 INFO [M:0;jenkins-hbase3:42279 {}] master.BackupLogCleaner(171): Stopping BackupLogCleaner 2023-08-10T06:33:25,341 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2023-08-10T06:33:25,341 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.large.0-1691649123785 {}] cleaner.HFileCleaner(306): Exit Thread[master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.large.0-1691649123785,5,FailOnTimeoutGroup] 2023-08-10T06:33:25,342 DEBUG [M:0;jenkins-hbase3:42279 {}] zookeeper.ReadOnlyZKClient(365): Close zookeeper connection 0x68964ef7 to 127.0.0.1:52413 2023-08-10T06:33:25,342 DEBUG [master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.small.0-1691649123785 {}] cleaner.HFileCleaner(306): Exit Thread[master/jenkins-hbase3:0:becomeActiveMaster-HFileCleaner.small.0-1691649123785,5,FailOnTimeoutGroup] 2023-08-10T06:33:25,342 DEBUG [M:0;jenkins-hbase3:42279 {}] ipc.AbstractRpcClient(513): Stopping rpc client 2023-08-10T06:33:25,342 DEBUG [M:0;jenkins-hbase3:42279 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2023-08-10T06:33:25,342 INFO [M:0;jenkins-hbase3:42279 {}] hbase.ChoreService(370): Chore service for: master/jenkins-hbase3:0 had [] on shutdown 2023-08-10T06:33:25,342 DEBUG [M:0;jenkins-hbase3:42279 {}] master.HMaster(1687): Stopping service threads 2023-08-10T06:33:25,342 INFO [M:0;jenkins-hbase3:42279 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2023-08-10T06:33:25,342 ERROR [M:0;jenkins-hbase3:42279 {}] procedure2.ProcedureExecutor(655): ThreadGroup java.lang.ThreadGroup[name=PEWorkerGroup,maxpri=10] contains running threads; null: See STDOUT java.lang.ThreadGroup[name=PEWorkerGroup,maxpri=10] Thread[HFileArchiver-2,5,PEWorkerGroup] 2023-08-10T06:33:25,342 INFO [M:0;jenkins-hbase3:42279 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2023-08-10T06:33:25,342 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2023-08-10T06:33:25,343 DEBUG [M:0;jenkins-hbase3:42279 {}] zookeeper.ZKUtil(347): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Unable to get data of znode /1/master because node does not exist (not an error) 2023-08-10T06:33:25,343 WARN [M:0;jenkins-hbase3:42279 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2023-08-10T06:33:25,343 INFO [M:0;jenkins-hbase3:42279 {}] assignment.AssignmentManager(382): Stopping assignment manager 2023-08-10T06:33:25,343 INFO [M:0;jenkins-hbase3:42279 {}] region.MasterRegion(178): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2023-08-10T06:33:25,344 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(1668): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2023-08-10T06:33:25,357 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(1690): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:33:25,357 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(1711): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:33:25,357 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(1778): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2023-08-10T06:33:25,357 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(1788): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:33:25,357 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(2819): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=203.11 KB heapSize=244.03 KB 2023-08-10T06:33:25,369 DEBUG [M:0;jenkins-hbase3:42279 {}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/135e5420c81f4c9cabdb7aa8624e3522 is 88, key is hbase:meta,,1/info:sn/1691649124781/Put/seqid=0 2023-08-10T06:33:25,400 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=None, state=Closed, path=null 2023-08-10T06:33:25,400 INFO [RS:0;jenkins-hbase3:34991 {}] regionserver.HRegionServer(1296): Exiting; stopping=jenkins-hbase3.apache.org,34991,1691649122119; zookeeper connection closed. 2023-08-10T06:33:25,400 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): regionserver:34991-0x10060ee705b0001, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=None, state=Closed, path=null 2023-08-10T06:33:25,400 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@77d95e4 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@77d95e4 2023-08-10T06:33:25,400 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2023-08-10T06:33:25,773 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=530 B at sequenceid=480 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/135e5420c81f4c9cabdb7aa8624e3522 2023-08-10T06:33:25,778 DEBUG [M:0;jenkins-hbase3:42279 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 135e5420c81f4c9cabdb7aa8624e3522 2023-08-10T06:33:25,792 DEBUG [M:0;jenkins-hbase3:42279 {}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a1616fb2f9934159983899b550df534d is 987, key is \x00\x00\x00\x00\x00\x00\x00\x1D/proc:d/1691649146084/Put/seqid=0 2023-08-10T06:33:26,198 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=202.51 KB at sequenceid=480 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a1616fb2f9934159983899b550df534d 2023-08-10T06:33:26,202 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.StoreFileReader(536): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for a1616fb2f9934159983899b550df534d 2023-08-10T06:33:26,203 DEBUG [M:0;jenkins-hbase3:42279 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: a1616fb2f9934159983899b550df534d 2023-08-10T06:33:26,215 DEBUG [M:0;jenkins-hbase3:42279 {}] hfile.HFileWriterImpl(813): Len of the biggest cell in hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/8c1fd1c064bd4cd58e035780d7f02189 is 82, key is jenkins-hbase3.apache.org,34991,1691649122119/rs:state/1691649123884/Put/seqid=0 2023-08-10T06:33:26,559 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(128): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2023-08-10T06:33:26,627 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=480 (bloomFilter=true), to=hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/8c1fd1c064bd4cd58e035780d7f02189 2023-08-10T06:33:26,631 DEBUG [M:0;jenkins-hbase3:42279 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 8c1fd1c064bd4cd58e035780d7f02189 2023-08-10T06:33:26,632 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/135e5420c81f4c9cabdb7aa8624e3522 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/135e5420c81f4c9cabdb7aa8624e3522 2023-08-10T06:33:26,637 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/135e5420c81f4c9cabdb7aa8624e3522, entries=8, sequenceid=480, filesize=5.5 K 2023-08-10T06:33:26,638 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/a1616fb2f9934159983899b550df534d as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/a1616fb2f9934159983899b550df534d 2023-08-10T06:33:26,643 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.StoreFileReader(536): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for a1616fb2f9934159983899b550df534d 2023-08-10T06:33:26,644 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/a1616fb2f9934159983899b550df534d, entries=53, sequenceid=480, filesize=16.7 K 2023-08-10T06:33:26,644 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.HRegionFileSystem(500): Committing hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/8c1fd1c064bd4cd58e035780d7f02189 as hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/8c1fd1c064bd4cd58e035780d7f02189 2023-08-10T06:33:26,649 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HStore$StoreFlusherImpl(1984): Added hdfs://localhost:33549/user/jenkins/test-data/7afc464b-130c-a783-5c09-f03fe2903cc3/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/8c1fd1c064bd4cd58e035780d7f02189, entries=1, sequenceid=480, filesize=5.1 K 2023-08-10T06:33:26,650 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(3022): Finished flush of dataSize ~203.11 KB/207981, heapSize ~243.73 KB/249584, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 1293ms, sequenceid=480, compaction requested=false 2023-08-10T06:33:26,660 DEBUG [StoreFileCloser-1595e783b53d99cd5eef43b6debb2682-info-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 135e5420c81f4c9cabdb7aa8624e3522 2023-08-10T06:33:26,662 DEBUG [StoreFileCloser-1595e783b53d99cd5eef43b6debb2682-proc-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: a1616fb2f9934159983899b550df534d 2023-08-10T06:33:26,676 DEBUG [StoreFileCloser-1595e783b53d99cd5eef43b6debb2682-rs-1 {}] hfile.PrefetchExecutor(137): Removing filename from the prefetched persistence list: 8c1fd1c064bd4cd58e035780d7f02189 2023-08-10T06:33:26,677 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(1909): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2023-08-10T06:33:26,677 DEBUG [M:0;jenkins-hbase3:42279 {}] regionserver.HRegion(1622): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2023-08-10T06:33:26,694 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2023-08-10T06:33:26,694 INFO [M:0;jenkins-hbase3:42279 {}] flush.MasterFlushTableProcedureManager(81): stop: server shutting down. 2023-08-10T06:33:26,694 INFO [M:0;jenkins-hbase3:42279 {}] master.LogRollMasterProcedureManager(73): stop: server shutting down. 2023-08-10T06:33:26,707 INFO [M:0;jenkins-hbase3:42279 {}] ipc.NettyRpcServer(213): Stopping server on /172.31.12.81:42279 2023-08-10T06:33:26,709 DEBUG [M:0;jenkins-hbase3:42279 {}] zookeeper.RecoverableZooKeeper(224): Node /1/rs/jenkins-hbase3.apache.org,42279,1691649119143 already deleted, retry=false 2023-08-10T06:33:26,824 INFO [M:0;jenkins-hbase3:42279 {}] regionserver.HRegionServer(1296): Exiting; stopping=jenkins-hbase3.apache.org,42279,1691649119143; zookeeper connection closed. 2023-08-10T06:33:26,824 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=None, state=Closed, path=null 2023-08-10T06:33:26,824 DEBUG [Listener at localhost/37579-EventThread {}] zookeeper.ZKWatcher(604): master:42279-0x10060ee705b0000, quorum=127.0.0.1:52413, baseZNode=/1 Received ZooKeeper Event, type=None, state=Closed, path=null 2023-08-10T06:33:26,827 WARN [Listener at jenkins-hbase3.apache.org/41523 {}] datanode.DirectoryScanner(537): DirectoryScanner: shutdown has been called 2023-08-10T06:33:26,829 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0 2023-08-10T06:33:26,938 WARN [BP-2138422956-172.31.12.81-1691649115962 heartbeating to localhost/127.0.0.1:33549 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2023-08-10T06:33:26,938 WARN [BP-2138422956-172.31.12.81-1691649115962 heartbeating to localhost/127.0.0.1:33549 {}] datanode.BPServiceActor(862): Ending block pool service for: Block pool BP-2138422956-172.31.12.81-1691649115962 (Datanode Uuid 400c1f01-c3a4-43fd-9e9d-b52c8cc54f17) service to localhost/127.0.0.1:33549 2023-08-10T06:33:26,939 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/cluster_31ee86f8-f5ed-8a45-aa31-8b420ad4a168/dfs/data/data1/current/BP-2138422956-172.31.12.81-1691649115962 {}] fs.CachingGetSpaceUsed$RefreshThread(211): Thread Interrupted waiting to refresh disk information: sleep interrupted 2023-08-10T06:33:26,939 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-backup/target/test-data/bd889c2c-aa2d-b062-e270-e3547eb38fd9/cluster_31ee86f8-f5ed-8a45-aa31-8b420ad4a168/dfs/data/data2/current/BP-2138422956-172.31.12.81-1691649115962 {}] fs.CachingGetSpaceUsed$RefreshThread(211): Thread Interrupted waiting to refresh disk information: sleep interrupted 2023-08-10T06:33:26,948 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0 2023-08-10T06:33:27,062 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2023-08-10T06:33:27,083 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2023-08-10T06:33:27,083 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.HBaseTestingUtility(2857): Stopping mini mapreduce cluster... 2023-08-10T06:33:27,095 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:0 2023-08-10T06:33:27,213 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:0 2023-08-10T06:33:27,327 ERROR [Thread[Thread-493,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(696): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2023-08-10T06:33:27,327 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:0 2023-08-10T06:33:27,434 WARN [ApplicationMaster Launcher {}] amlauncher.ApplicationMasterLauncher$LauncherThread(122): org.apache.hadoop.yarn.server.resourcemanager.amlauncher.ApplicationMasterLauncher$LauncherThread interrupted. Returning. 2023-08-10T06:33:27,442 ERROR [SchedulerEventDispatcher:Event Processor {}] event.EventDispatcher$EventProcessor(61): Returning, interrupted : java.lang.InterruptedException 2023-08-10T06:33:27,444 ERROR [Thread[Thread-503,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(696): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2023-08-10T06:33:27,454 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] log.Slf4jLog(67): Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@jenkins-hbase3.apache.org:0 2023-08-10T06:33:27,560 ERROR [Thread[Thread-468,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(696): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2023-08-10T06:33:27,560 INFO [Listener at jenkins-hbase3.apache.org/41523 {}] hbase.HBaseTestingUtility(2860): Mini mapreduce cluster stopped