Partilhar via


rx_spark_disconnect

Usage

revoscalepy.rx_spark_disconnect(compute_context=None)

Description

Shuts down the remote Spark application and switches to a local compute context. All rx* function calls after this will run in a local compute context. In pyspark-interop mode, if Spark application is started by pyspark APIs, rx_spark_disconnect will not shut down the remote Spark application but disassociate from it. Run ‘help(revoscalepy.rx_spark_connect)’ for more information about interop.

Arguments

compute_context

Spark compute context to be terminated by rx_spark_disconnect. If input is None, then current compute context will be used.

Example

from revoscalepy import rx_spark_connect, rx_spark_disconnect
cc = rx_spark_connect()
rx_spark_disconnect(cc)