This documentation is for an out-of-date version of Apache Flink. We recommend you use the latest stable version.

Memory tuning guide

In addition to the main memory setup guide, this section explains how to set up memory depending on the use case and which options are important for each case.

Configure memory for standalone deployment

It is recommended to configure total Flink memory (taskmanager.memory.flink.size or jobmanager.memory.flink.size) or its components for standalone deployment where you want to declare how much memory is given to Flink itself. Additionally, you can adjust JVM metaspace if it causes problems.

The total Process memory is not relevant because JVM overhead is not controlled by Flink or the deployment environment, only physical resources of the executing machine matter in this case.

Configure memory for containers

It is recommended to configure total process memory (taskmanager.memory.process.size or jobmanager.memory.process.size) for the containerized deployments (Kubernetes, Yarn or Mesos). It declares how much memory in total should be assigned to the Flink JVM process and corresponds to the size of the requested container.

Note If you configure the total Flink memory Flink will implicitly add JVM memory components to derive the total process memory and request a container with the memory of that derived size.

Warning: If Flink or user code allocates unmanaged off-heap (native) memory beyond the container size the job can fail because the deployment environment can kill the offending containers.

See also description of container memory exceeded failure.

Configure memory for state backends

This is only relevant for TaskManagers.

When deploying a Flink streaming application, the type of state backend used will dictate the optimal memory configurations of your cluster.

Heap state backend

When running a stateless job or using a heap state backend (MemoryStateBackend or FsStateBackend), set managed memory to zero. This will ensure that the maximum amount of heap memory is allocated for user code on the JVM.

RocksDB state backend

The RocksDBStateBackend uses native memory. By default, RocksDB is set up to limit native memory allocation to the size of the managed memory. Therefore, it is important to reserve enough managed memory for your state. If you disable the default RocksDB memory control, TaskManagers can be killed in containerized deployments if RocksDB allocates memory above the limit of the requested container size (the total process memory). See also how to tune RocksDB memory and state.backend.rocksdb.memory.managed.

Configure memory for batch jobs

This is only relevant for TaskManagers.

Flink’s batch operators leverage managed memory to run more efficiently. In doing so, some operations can be performed directly on raw data without having to be deserialized into Java objects. This means that managed memory configurations have practical effects on the performance of your applications. Flink will attempt to allocate and use as much managed memory as configured for batch jobs but not go beyond its limits. This prevents OutOfMemoryError’s because Flink knows precisely how much memory it has to leverage. If the managed memory is not sufficient, Flink will gracefully spill to disk.

Configure memory for sort-merge blocking shuffle

The number of required network buffers per sort-merge blocking result partition is controlled by and the default value is 64 which is quite small. Though it can work for arbitrary parallelism, the performance may not be the best. For large scale jobs, it is suggested to increase this config value to improve compression ratio and reduce small network packets which is good for performance. To increase this value, you may also need to increase the size of total network memory by adjusting the config values of, and taskmanager. to avoid insufficient number of network buffers error.

Except for network memory, the sort-merge blocking shuffle implementation also uses some unmanaged direct memory for shuffle data writing and reading. So to use sort-merge blocking shuffle, you may need to reserve some direct memory for it by increasing the config value of taskmanager.memory.task .off-heap.size. If direct memory OOM error occurs after you enable the sort-merge blocking shuffle, you can just give more direct memory until the OOM error disappears.