| 
Deprecated Methods | 
org.apache.spark.streaming.StreamingContext.awaitTermination(long)
 
          As of 1.3.0, replaced by awaitTerminationOrTimeout(Long).  | 
org.apache.spark.streaming.api.java.JavaStreamingContext.awaitTermination(long)
 
          As of 1.3.0, replaced by awaitTerminationOrTimeout(Long).  | 
org.apache.spark.api.java.StorageLevels.create(boolean, boolean, boolean, int)
 
            | 
org.apache.spark.sql.DataFrame.createJDBCTable(String, String, boolean)
 
          As of 1.340, replaced by write().jdbc().  | 
org.apache.spark.api.java.JavaSparkContext.defaultMinSplits()
 
          As of Spark 1.0.0, defaultMinSplits is deprecated, use
            JavaSparkContext.defaultMinPartitions() instead  | 
org.apache.spark.streaming.api.java.JavaDStreamLike.foreach(Function)
 
          As of release 0.9.0, replaced by foreachRDD  | 
org.apache.spark.streaming.dstream.DStream.foreach(Function1, BoxedUnit>)
 
          As of 0.9.0, replaced by foreachRDD.  | 
org.apache.spark.streaming.dstream.DStream.foreach(Function2, Time, BoxedUnit>)
 
          As of 0.9.0, replaced by foreachRDD.  | 
org.apache.spark.streaming.api.java.JavaDStreamLike.foreach(Function2)
 
          As of release 0.9.0, replaced by foreachRDD  | 
org.apache.spark.sql.types.DataType.fromCaseClassString(String)
 
          As of 1.2.0, replaced by DataType.fromJson()  | 
org.apache.spark.streaming.api.java.JavaStreamingContext.getOrCreate(String, Configuration, JavaStreamingContextFactory)
 
          As of 1.4.0, replaced by getOrCreate without JavaStreamingContextFactor.  | 
org.apache.spark.streaming.api.java.JavaStreamingContext.getOrCreate(String, Configuration, JavaStreamingContextFactory, boolean)
 
          As of 1.4.0, replaced by getOrCreate without JavaStreamingContextFactor.  | 
org.apache.spark.streaming.api.java.JavaStreamingContext.getOrCreate(String, JavaStreamingContextFactory)
 
          As of 1.4.0, replaced by getOrCreate without JavaStreamingContextFactor.  | 
org.apache.spark.sql.DataFrame.insertInto(String)
 
          As of 1.4.0, replaced by
            write().mode(SaveMode.Append).saveAsTable(tableName).  | 
org.apache.spark.sql.DataFrame.insertInto(String, boolean)
 
          As of 1.4.0, replaced by
            write().mode(SaveMode.Append|SaveMode.Overwrite).saveAsTable(tableName).  | 
org.apache.spark.sql.DataFrame.insertIntoJDBC(String, String, boolean)
 
          As of 1.4.0, replaced by write().jdbc().  | 
org.apache.spark.sql.SQLContext.jdbc(String, String)
 
          As of 1.4.0, replaced by read().jdbc().  | 
org.apache.spark.sql.SQLContext.jdbc(String, String, String[])
 
          As of 1.4.0, replaced by read().jdbc().  | 
org.apache.spark.sql.SQLContext.jdbc(String, String, String, long, long, int)
 
          As of 1.4.0, replaced by read().jdbc().  | 
org.apache.spark.sql.SQLContext.jsonFile(String)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.jsonFile(String, double)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.jsonFile(String, StructType)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD, double)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD, StructType)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.jsonRDD(RDD)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.jsonRDD(RDD, double)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.jsonRDD(RDD, StructType)
 
          As of 1.4.0, replaced by read().json().  | 
org.apache.spark.sql.SQLContext.load(String)
 
          As of 1.4.0, replaced by read().load(path).  | 
org.apache.spark.sql.SQLContext.load(String, Map)
 
          As of 1.4.0, replaced by read().format(source).options(options).load().  | 
org.apache.spark.sql.SQLContext.load(String, Map)
 
          As of 1.4.0, replaced by read().format(source).options(options).load().  | 
org.apache.spark.sql.SQLContext.load(String, String)
 
          As of 1.4.0, replaced by read().format(source).load(path).  | 
org.apache.spark.sql.SQLContext.load(String, StructType, Map)
 
          As of 1.4.0, replaced by
            read().format(source).schema(schema).options(options).load().  | 
org.apache.spark.sql.SQLContext.load(String, StructType, Map)
 
          As of 1.4.0, replaced by
            read().format(source).schema(schema).options(options).load().  | 
org.apache.spark.mllib.util.MLUtils.loadLabeledData(SparkContext, String)
 
          Should use RDD.saveAsTextFile(java.lang.String) for saving and
            MLUtils.loadLabeledPoints(org.apache.spark.SparkContext, java.lang.String, int) for loading.  | 
org.apache.spark.streaming.StreamingContext.networkStream(Receiver, ClassTag)
 
          As of 1.0.0", replaced by receiverStream.  | 
org.apache.spark.sql.SQLContext.parquetFile(String...)
 
          As of 1.4.0, replaced by read().parquet().  | 
org.apache.spark.streaming.api.java.JavaDStreamLike.reduceByWindow(Function2, Duration, Duration)
 
          As this API is not Java compatible.  | 
org.apache.spark.sql.DataFrame.save(String)
 
          As of 1.4.0, replaced by write().save(path).  | 
org.apache.spark.sql.DataFrame.save(String, SaveMode)
 
          As of 1.4.0, replaced by write().mode(mode).save(path).  | 
org.apache.spark.sql.DataFrame.save(String, SaveMode, Map)
 
          As of 1.4.0, replaced by
            write().format(source).mode(mode).options(options).save(path).  | 
org.apache.spark.sql.DataFrame.save(String, SaveMode, Map)
 
          As of 1.4.0, replaced by
            write().format(source).mode(mode).options(options).save(path).  | 
org.apache.spark.sql.DataFrame.save(String, String)
 
          As of 1.4.0, replaced by write().format(source).save(path).  | 
org.apache.spark.sql.DataFrame.save(String, String, SaveMode)
 
          As of 1.4.0, replaced by write().format(source).mode(mode).save(path).  | 
org.apache.spark.sql.DataFrame.saveAsParquetFile(String)
 
          As of 1.4.0, replaced by write().parquet().  | 
org.apache.spark.sql.DataFrame.saveAsTable(String)
 
          As of 1.4.0, replaced by write().saveAsTable(tableName).  | 
org.apache.spark.sql.DataFrame.saveAsTable(String, SaveMode)
 
          As of 1.4.0, replaced by write().mode(mode).saveAsTable(tableName).  | 
org.apache.spark.sql.DataFrame.saveAsTable(String, String)
 
          As of 1.4.0, replaced by write().format(source).saveAsTable(tableName).  | 
org.apache.spark.sql.DataFrame.saveAsTable(String, String, SaveMode)
 
          As of 1.4.0, replaced by write().mode(mode).saveAsTable(tableName).  | 
org.apache.spark.sql.DataFrame.saveAsTable(String, String, SaveMode, Map)
 
          As of 1.4.0, replaced by
            write().format(source).mode(mode).options(options).saveAsTable(tableName).  | 
org.apache.spark.sql.DataFrame.saveAsTable(String, String, SaveMode, Map)
 
          As of 1.4.0, replaced by
            write().format(source).mode(mode).options(options).saveAsTable(tableName).  | 
org.apache.spark.mllib.util.MLUtils.saveLabeledData(RDD, String)
 
          Should use RDD.saveAsTextFile(java.lang.String) for saving and
            MLUtils.loadLabeledPoints(org.apache.spark.SparkContext, java.lang.String, int) for loading.  | 
org.apache.spark.streaming.api.java.JavaStreamingContext.sc()
 
          As of 0.9.0, replaced by sparkContext  | 
org.apache.spark.mllib.optimization.LBFGS.setMaxNumIterations(int)
 
          use LBFGS.setNumIterations(int) instead  | 
org.apache.spark.api.java.JavaRDDLike.toArray()
 
          As of Spark 1.0.0, toArray() is deprecated, use JavaRDDLike.collect() instead  | 
org.apache.spark.streaming.StreamingContext.toPairDStreamFunctions(DStream>, ClassTag, ClassTag, Ordering)
 
          As of 1.3.0, replaced by implicit functions in the DStream companion object.
             This is kept here only for backward compatibility.  | 
org.apache.spark.sql.DataFrame.toSchemaRDD()
 
          As of 1.3.0, replaced by toDF().  | 
org.apache.spark.mllib.rdd.RDDFunctions.treeAggregate(U, Function2, Function2, int, ClassTag)
 
          Use RDD.treeAggregate(U, scala.Function2, scala.Function2, int, scala.reflect.ClassTag) instead.  | 
org.apache.spark.mllib.rdd.RDDFunctions.treeReduce(Function2, int)
 
          Use RDD.treeReduce(scala.Function2, int) instead.  |