Rescale Bucket

Rescale Bucket #

Since the number of total buckets dramatically influences the performance, Table Store allows users to tune bucket numbers by ALTER TABLE command and reorganize data layout by INSERT OVERWRITE without recreating the table/partition. When executing overwrite jobs, the framework will automatically scan the data with the old bucket number and hash the record according to the current bucket number.

Rescale Overwrite #

-- rescale number of total buckets
ALTER TABLE table_identifier SET ('bucket' = '...')

-- reorganize data layout of table/partition
INSERT OVERWRITE table_identifier [PARTITION (part_spec)]
SELECT ... 
FROM table_identifier
[WHERE part_spec]

Please note that

  • ALTER TABLE only modifies the table’s metadata and will NOT reorganize or reformat existing data. Reorganize exiting data must be achieved by INSERT OVERWRITE.
  • Rescale bucket number does not influence the read and running write jobs.
  • Once the bucket number is changed, any newly scheduled INSERT INTO jobs which write to without-reorganized existing table/partition will throw a TableException with message like
    Try to write table/partition ... with a new bucket num ..., 
    but the previous bucket num is ... Please switch to batch mode, 
    and perform INSERT OVERWRITE to rescale current data layout first.
    
  • For partitioned table, it is possible to have different bucket number for different partitions. E.g.
    ALTER TABLE my_table SET ('bucket' = '4');
    INSERT OVERWRITE my_table PARTITION (dt = '2022-01-01')
    SELECT * FROM ...;
      
    ALTER TABLE my_table SET ('bucket' = '8');
    INSERT OVERWRITE my_table PARTITION (dt = '2022-01-02')
    SELECT * FROM ...;
    
  • During overwrite period, make sure there are no other jobs writing the same table/partition.
Note: For the table which enables log system(e.g. Kafka), please rescale the topic’s partition as well to keep consistency.

Use Case #

Rescale bucket helps to handle sudden spikes in throughput. Suppose there is a daily streaming ETL task to sync transaction data. The table’s DDL and pipeline are listed as follows.

-- table DDL
CREATE TABLE verified_orders (
    trade_order_id BIGINT,
    item_id BIGINT,
    item_price DOUBLE,
    dt STRING,
    PRIMARY KEY (dt, trade_order_id, item_id) NOT ENFORCED 
) PARTITIONED BY (dt)
WITH (
    'bucket' = '16'
);

-- streaming insert as bucket num = 16
INSERT INTO verified_orders
SELECT trade_order_id,
       item_id,
       item_price,
       DATE_FORMAT(gmt_create, 'yyyy-MM-dd') AS dt
FROM raw_orders
WHERE order_status = 'verified'

The pipeline has been running well for the past few weeks. However, the data volume has grown fast recently, and the job’s latency keeps increasing. To improve the data freshness, users can

  • Suspend the streaming job with a savepoint ( see Suspended State and Stopping a Job Gracefully Creating a Final Savepoint )
    $ ./bin/flink stop \
          --savepointPath /tmp/flink-savepoints \
          $JOB_ID
    
  • Increase the bucket number
    -- scaling out
    ALTER TABLE verified_orders SET ('bucket' = '32')
    
  • Switch to the batch mode and overwrite the current partition(s) to which the streaming job is writing
    SET 'execution.runtime-mode' = 'batch';
    -- suppose today is 2022-06-22
    -- case 1: there is no late event which updates the historical partitions, thus overwrite today's partition is enough
    INSERT OVERWRITE verified_orders PARTITION (dt = '2022-06-22')
    SELECT trade_order_id,
           item_id,
           item_price
    FROM verified_orders
    WHERE dt = '2022-06-22' AND order_status = 'verified'
      
    -- case 2: there are late events updating the historical partitions, but the range does not exceed 3 days
    INSERT OVERWRITE verified_orders
    SELECT trade_order_id,
           item_id,
           item_price,
           dt
    FROM verified_orders
    WHERE dt IN ('2022-06-20', '2022-06-21', '2022-06-22') AND order_status = 'verified'
    
  • After overwrite job finished, switch back to streaming mode. And now, the parallelism can be increased alongside with bucket number to restore the streaming job from the savepoint ( see Start a SQL Job from a savepoint )
    SET 'execution.runtime-mode' = 'streaming';
    SET 'execution.savepoint.path' = <savepointPath>;
    
    INSERT INTO verified_orders
    SELECT trade_order_id,
         item_id,
         item_price,
         DATE_FORMAT(gmt_create, 'yyyy-MM-dd') AS dt
    FROM raw_orders
    WHERE order_status = 'verified'