Defines operations common to several Java RDD implementations.
Defines operations common to several Java RDD implementations. Note that this trait is not intended to be implemented by user code.
A Java-friendly version of org.apache.spark.SparkContext that returns org.apache.spark.api.java.JavaRDDs and works with Java collections instead of Scala ones.
A Java-friendly version of org.apache.spark.SparkContext that returns org.apache.spark.api.java.JavaRDDs and works with Java collections instead of Scala ones.
Only one SparkContext may be active per JVM. You must stop()
the active SparkContext before
creating a new one. This limitation may eventually be removed; see SPARK-2243 for more details.
Low-level status reporting APIs for monitoring job and stage progress.
Low-level status reporting APIs for monitoring job and stage progress.
These APIs intentionally provide very weak consistency semantics; consumers of these APIs should
be prepared to handle empty / missing information. For example, a job's stage ids may be known
but the status API may not have any information about the details of those stages, so
getStageInfo
could potentially return null
for a valid stage id.
To limit memory usage, these APIs only provide information on recent jobs / stages. These APIs
will provide information for the last spark.ui.retainedStages
stages and
spark.ui.retainedJobs
jobs.
NOTE: this class's constructor should be considered private and may be subject to change.
Set of interfaces to represent functions in Spark's Java API.
Set of interfaces to represent functions in Spark's Java API. Users create implementations of these interfaces to pass functions to various Java API methods for Spark. Please visit Spark's Java programming guide for more details.
Spark Java programming APIs.