Packages

  • package root
    Definition Classes
    root
  • package org
    Definition Classes
    root
  • package apache
    Definition Classes
    org
  • package flink
    Definition Classes
    apache
  • package api
    Definition Classes
    flink
  • package scala

    The Flink Scala API.

    The Flink Scala API. org.apache.flink.api.scala.ExecutionEnvironment is the starting-point of any Flink program. It can be used to read from local files, HDFS, or other sources. org.apache.flink.api.scala.DataSet is the main abstraction of data in Flink. It provides operations that create new DataSets via transformations. org.apache.flink.api.scala.GroupedDataSet provides operations on grouped data that results from org.apache.flink.api.scala.DataSet.groupBy().

    Use org.apache.flink.api.scala.ExecutionEnvironment.getExecutionEnvironment to obtain an execution environment. This will either create a local environment or a remote environment, depending on the context where your program is executing.

    Definition Classes
    api
    Deprecated

    All Flink Scala APIs are deprecated and will be removed in a future Flink version. You can still build your application in Scala, but you should move to the Java version of either the DataStream and/or Table API.

    See also

    FLIP-265 Deprecate and remove Scala API support

  • package extensions

    acceptPartialFunctions extends the original DataSet with methods with unique names that delegate to core higher-order functions (e.g.

    acceptPartialFunctions extends the original DataSet with methods with unique names that delegate to core higher-order functions (e.g. map) so that we can work around the fact that overloaded methods taking functions as parameters can't accept partial functions as well. This enables the possibility to directly apply pattern matching to decompose inputs such as tuples, case classes and collections.

    The following is a small example that showcases how this extensions would work on a Flink data set:

    object Main {
      import org.apache.flink.api.scala.extensions._
      case class Point(x: Double, y: Double)
      def main(args: Array[String]): Unit = {
        val env = ExecutionEnvironment.getExecutionEnvironment
        val ds = env.fromElements(Point(1, 2), Point(3, 4), Point(5, 6))
        ds.filterWith {
          case Point(x, _) => x > 1
        }.reduceWith {
          case (Point(x1, y1), (Point(x2, y2))) => Point(x1 + y1, x2 + y2)
        }.mapWith {
          case Point(x, y) => (x, y)
        }.flatMapWith {
          case (x, y) => Seq('x' -> x, 'y' -> y)
        }.groupingBy {
          case (id, value) => id
        }
      }
    }

    The extension consists of several implicit conversions over all the data set representations that could gain from this feature. To use this set of extensions methods the user has to explicitly opt-in by importing org.apache.flink.api.scala.extensions.acceptPartialFunctions.

    For more information and usage examples please consult the Apache Flink official documentation.

    Definition Classes
    scala
    Deprecated

    All Flink Scala APIs are deprecated and will be removed in a future Flink major version. You can still build your application in Scala, but you should move to the Java version of either the DataStream and/or Table API.

    See also

    FLIP-265 Deprecate and remove Scala API support

  • package impl
    Definition Classes
    extensions
  • package acceptPartialFunctions
    Definition Classes
    impl
  • OnCoGroupDataSet
  • OnCrossDataSet
  • OnDataSet
  • OnGroupedDataSet
  • OnHalfUnfinishedKeyPairOperation
  • OnJoinFunctionAssigner
  • OnUnfinishedKeyPairOperation

class OnGroupedDataSet[T] extends AnyRef

Wraps a grouped data set, allowing to use anonymous partial functions to perform extraction of items in a tuple, case class instance or collection

T

The type of the grouped data set items

Annotations
@deprecated
Deprecated

(Since version 1.18.0)

See also

FLIP-265 Deprecate and remove Scala API support

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. OnGroupedDataSet
  2. AnyRef
  3. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new OnGroupedDataSet(ds: GroupedDataSet[T])

    ds

    The wrapped grouped data set

    Deprecated

    All Flink Scala APIs are deprecated and will be removed in a future Flink major version. You can still build your application in Scala, but you should move to the Java version of either the DataStream and/or Table API.

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @native() @throws( ... )
  6. def combineGroupWith[R](fun: (Stream[T]) ⇒ R)(implicit arg0: TypeInformation[R], arg1: ClassTag[R]): DataSet[R]

    Same as a reducing operation but only acts locally, ideal to perform pre-aggregation before a reduction.

    Same as a reducing operation but only acts locally, ideal to perform pre-aggregation before a reduction.

    R

    The type of the items in the resulting data set

    fun

    The reducing function

    returns

    A data set of Rs reduced group-wise

    Annotations
    @PublicEvolving()
  7. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  8. def equals(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  9. def finalize(): Unit
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  10. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  11. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  12. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  13. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  14. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  15. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  16. def reduceGroupWith[R](fun: (Stream[T]) ⇒ R)(implicit arg0: TypeInformation[R], arg1: ClassTag[R]): DataSet[R]

    Reduces the data set group-wise with a reducer fun

    Reduces the data set group-wise with a reducer fun

    R

    The type of the items in the resulting data set

    fun

    The reducing function

    returns

    A data set of Rs reduced group-wise

    Annotations
    @PublicEvolving()
  17. def reduceWith(fun: (T, T) ⇒ T): DataSet[T]

    Reduces the whole data set with a reducer fun

    Reduces the whole data set with a reducer fun

    fun

    The reducing function

    returns

    A reduced data set of Ts

    Annotations
    @PublicEvolving()
  18. def sortGroupWith[K](order: Order)(fun: (T) ⇒ K)(implicit arg0: TypeInformation[K]): GroupedDataSet[T]

    Sorts a group using a sorting function fun and an Order

    Sorts a group using a sorting function fun and an Order

    K

    The key type

    order

    The ordering strategy (ascending, descending, etc.)

    fun

    The sorting function, defining the sorting key

    returns

    A data set sorted group-wise

    Annotations
    @PublicEvolving()
  19. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  20. def toString(): String
    Definition Classes
    AnyRef → Any
  21. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  22. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  23. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @throws( ... )

Inherited from AnyRef

Inherited from Any

Ungrouped