About 9,200,000 results
Open links in new tab
  1. scala - What is RDD in spark - Stack Overflow

    Dec 23, 2015 · An RDD is, essentially, the Spark representation of a set of data, spread across multiple machines, with APIs to let you act on it. An RDD could come from any datasource, …

  2. Difference between DataFrame, Dataset, and RDD in Spark

    Feb 18, 2020 · I'm just wondering what is the difference between an RDD and DataFrame (Spark 2.0.0 DataFrame is a mere type alias for Dataset[Row]) in Apache Spark? Can you convert …

  3. Difference between Spark RDDs and HDFS' data blocks

    Jan 31, 2018 · Is there any relation to HDFS' data blocks? In general not. They address different issues RDDs are about distributing computation and handling computation failures. HDFS is …

  4. How do I split an RDD into two or more RDDs? - Stack Overflow

    Oct 6, 2015 · I'm looking for a way to split an RDD into two or more RDDs. The closest I've seen is Scala Spark: Split collection into several RDD? which is still a single RDD. If you're familiar …

  5. hadoop - What is Lineage In Spark? - Stack Overflow

    Aug 18, 2017 · In Spark, Lineage Graph is a dependencies graph in between existing RDD and new RDD. It means that all the dependencies between the RDD will be recorded in a graph, …

  6. What is the difference between spark checkpoint and persist to a …

    Feb 1, 2016 · RDD checkpointing is a different concept than a chekpointing in Spark Streaming. The former one is designed to address lineage issue, the latter one is all about streaming …

  7. What's the difference between RDD and Dataframe in Spark?

    Aug 20, 2019 · RDD stands for Resilient Distributed Datasets. It is Read-only partition collection of records. RDD is the fundamental data structure of Spark. It allows a programmer to perform …

  8. View RDD contents in Python Spark? - Stack Overflow

    Please note that when you run collect (), the RDD - which is a distributed data set is aggregated at the driver node and is essentially converted to a list. So obviously, it won't be a good idea to …

  9. Splitting an Pyspark RDD into Different columns and convert to …

    How do I split and convert the RDD to Dataframe in pyspark such that, the first element is taken as first column, and the rest elements combined to a single column ?

  10. How to find an average for a Spark RDD? - Stack Overflow

    Jul 9, 2018 · rdd.reduce ( (_ + _) / 2) There are a few issues with the above reduce method for average calculation: The placeholder syntax won't work as the shorthand for reduce((acc, x) …