Packages

  • package root
    Definition Classes
    root
  • package org
    Definition Classes
    root
  • package apache
    Definition Classes
    org
  • package flink
    Definition Classes
    apache
  • package api
    Definition Classes
    flink
  • package scala

    The Flink Scala API.

    The Flink Scala API. org.apache.flink.api.scala.ExecutionEnvironment is the starting-point of any Flink program. It can be used to read from local files, HDFS, or other sources. org.apache.flink.api.scala.DataSet is the main abstraction of data in Flink. It provides operations that create new DataSets via transformations. org.apache.flink.api.scala.GroupedDataSet provides operations on grouped data that results from org.apache.flink.api.scala.DataSet.groupBy().

    Use org.apache.flink.api.scala.ExecutionEnvironment.getExecutionEnvironment to obtain an execution environment. This will either create a local environment or a remote environment, depending on the context where your program is executing.

    Definition Classes
    api
    Deprecated

    All Flink Scala APIs are deprecated and will be removed in a future Flink version. You can still build your application in Scala, but you should move to the Java version of either the DataStream and/or Table API.

    See also

    FLIP-265 Deprecate and remove Scala API support

  • package extensions

    acceptPartialFunctions extends the original DataSet with methods with unique names that delegate to core higher-order functions (e.g.

    acceptPartialFunctions extends the original DataSet with methods with unique names that delegate to core higher-order functions (e.g. map) so that we can work around the fact that overloaded methods taking functions as parameters can't accept partial functions as well. This enables the possibility to directly apply pattern matching to decompose inputs such as tuples, case classes and collections.

    The following is a small example that showcases how this extensions would work on a Flink data set:

    object Main {
      import org.apache.flink.api.scala.extensions._
      case class Point(x: Double, y: Double)
      def main(args: Array[String]): Unit = {
        val env = ExecutionEnvironment.getExecutionEnvironment
        val ds = env.fromElements(Point(1, 2), Point(3, 4), Point(5, 6))
        ds.filterWith {
          case Point(x, _) => x > 1
        }.reduceWith {
          case (Point(x1, y1), (Point(x2, y2))) => Point(x1 + y1, x2 + y2)
        }.mapWith {
          case Point(x, y) => (x, y)
        }.flatMapWith {
          case (x, y) => Seq('x' -> x, 'y' -> y)
        }.groupingBy {
          case (id, value) => id
        }
      }
    }

    The extension consists of several implicit conversions over all the data set representations that could gain from this feature. To use this set of extensions methods the user has to explicitly opt-in by importing org.apache.flink.api.scala.extensions.acceptPartialFunctions.

    For more information and usage examples please consult the Apache Flink official documentation.

    Definition Classes
    scala
    Deprecated

    All Flink Scala APIs are deprecated and will be removed in a future Flink major version. You can still build your application in Scala, but you should move to the Java version of either the DataStream and/or Table API.

    See also

    FLIP-265 Deprecate and remove Scala API support

  • package metrics
    Definition Classes
    scala
  • package operators
    Definition Classes
    scala
  • package typeutils
    Definition Classes
    scala
  • package utils
    Definition Classes
    scala
  • DataSetUtils

package utils

Linear Supertypes
AnyRef, Any

Type Members

  1. implicit class DataSetUtils[T] extends AnyRef

    This class provides simple utility methods for zipping elements in a data set with an index or with a unique identifier, sampling elements from a data set.

    This class provides simple utility methods for zipping elements in a data set with an index or with a unique identifier, sampling elements from a data set.

    Annotations
    @PublicEvolving()
    Deprecated

    All Flink Scala APIs are deprecated and will be removed in a future Flink major version. You can still build your application in Scala, but you should move to the Java version of either the DataStream and/or Table API.

    See also

    FLIP-265 Deprecate and remove Scala API support

Inherited from AnyRef

Inherited from Any

Ungrouped