Finding the average of a data set using Apache Spark - scala
I am learning how to use Apache Spark and I am trying to get the average temperature from each hour from a data set. The data set that I am trying to use is from weather information stored in a csv. I am having trouble finding how to first read in the csv file and then calculating the average temperature for each hour.
From the spark documentation I am using the example Scala line to read in a file.
val textFile = sc.textFile("README.md")
I have given the link for the data file below. I am using the file called JCMB_2014.csv as it is the latest one with all months covered.
Weather Data
Edit:
The code I have tried so far is:
class SimpleCSVHeader(header:Array[String]) extends Serializable {
val index = header.zipWithIndex.toMap
def apply(array:Array[String], key:String):String = array(index(key))
}
val csv = sc.textFile("JCMB_2014.csv")
val data = csv.map(line => line.split(",").map(elem => elem.trim))
val header = new SimpleCSVHeader(data.take(1)(0)) // we build our header
val header = new SimpleCSVHeader(data.take(1)(0))
val rows = data.filter(line => header(line,"date-time") != "date-time")
val users = rows.map(row => header(row,"date-time")
val usersByHits = rows.map(row => header(row,"date-time") -> header(row,"surface temperature (C)").toInt)
Here is sample code for calculating averages on hourly basis
Step1:Read file, Filter header,extract time and temp columns
scala> val hourlyTemps = lines.map(line=>line.split(",")).filter(entries=>(!"time".equals(entries(3)))).map(entries=>(entries(3).toInt/60,(entries(8).toFloat,1)))
scala> hourlyTemps.take(1)
res25: Array[(Int, (Float, Int))] = Array((9,(10.23,1)))
(time/60) discards minutes and keeps only hours
Step2:Aggregate temperatures and no of occurrences
scala> val aggregateTemps=hourlyTemps.reduceByKey((a,b)=>(a._1+b._1,a._2+b._2))
scala> aggreateTemps.take(1)
res26: Array[(Int, (Double, Int))] = Array((34,(8565.25,620)))
Step2:Calculate Averages using total and no of occurrences
Find the final result below.
val avgTemps=aggregateTemps.map(tuple=>(tuple._1,tuple._2._1/tuple._2._2))
scala> avgTemps.collect
res28: Array[(Int, Float)] = Array((34,13.814922), (4,11.743354), (16,14.227251), (22,15.770312), (28,15.5324545), (30,15.167026), (14,13.177828), (32,14.659948), (36,12.865237), (0,11.994799), (24,15.662579), (40,12.040322), (6,11.398838), (8,11.141323), (12,12.004652), (38,12.329914), (18,15.020147), (20,15.358524), (26,15.631921), (10,11.192643), (2,11.848178), (13,12.616284), (19,15.198371), (39,12.107664), (15,13.706351), (21,15.612191), (25,15.627121), (29,15.432097), (11,11.541124), (35,13.317129), (27,15.602408), (33,14.220147), (37,12.644306), (23,15.83412), (1,11.872819), (17,14.595772), (3,11.78971), (7,11.248139), (9,11.049844), (31,14.901464), (5,11.59693))
You may want to provide Structure definition of your CSV file and convert your RDD to DataFrame, like described in the documentation. Dataframes provide a whole set of useful predefined statistic functions as well as the possibility to write some simple custom functions. You then will be able to compute the average with:
dataFrame.groupBy(<your columns here>).agg(avg(<column to compute average>)
Related
How to filter an rdd by data type?
I have an rdd that i am trying to filter for only float type. Do Spark rdds provide any way of doing this? I have a csv where I need only float values greater than 40 into a new rdd. To achieve this, i am checking if it is an instance of type float and filtering them. When I filter with a !, all the strings are still there in the output and when i dont use !, the output is empty. val airports1 = airports.filter(line => !line.split(",")(6).isInstanceOf[Float]) val airports2 = airports1.filter(line => line.split(",")(6).toFloat > 40) At the .toFloat , i run into NumberFormatException which I've tried to handle in a try catch block.
Since you have a plain string and you are trying to get float values from it, you are not actually filtering by type. But, if they can be parsed to float instead. You can accomplish that using a flatMap together with Option. import org.apache.spark.sql.SparkSession import scala.util.Try val spark = SparkSession.builder.master("local[*]").appName("Float caster").getOrCreate() val sc = spark.sparkContext val data = List("x,10", "y,3.3", "z,a") val rdd = sc.parallelize(data) // rdd: RDD[String] val filtered = rdd.flatMap(line => Try(line.split(",")(1).toFloat).toOption) // filtered: RDD[Float] filtered.collect() // res0: Array[Float] = Array(10.0, 3.3) For the > 40 part you can either, perform another filter after or filter the inner Option. (Both should perform more or less equals due spark laziness, thus choose the one is more clear for you). // Option 1 - Another filter. val filtered2 = filtered.filter(x => x > 40) // Option 2 - Filter the inner option in one step. val filtered = rdd.flatMap(line => Try(line.split(",")(1).toFloat).toOption.filter(x => x > 40)) Let me know if you have any question.
Scala - How to read a csv table into a RDD[Vector]
I would like to read from a huge csv file, assign every row to a vector via spliting values by ",". In the end I aim to have an RDD of Vectors which holds the values. However I get an error after Seq: type mismatch; found : Unit required: org.apache.spark.mllib.linalg.Vector Error occurred in an application involving default arguments. My code is like this so far: val file = "/data.csv" val data: RDD[Vector] =sc.parallelize( Seq( for(line <- Source.fromFile(file).getLines){ Vectors.dense(line.split (",").map (_.toDouble).distinct) } ) )
You should read it using sparkContext's textFile api as val file = "/data.csv" val data = sc.textFile(file).map(line => Vectors.dense(line.split (",").map(_.toDouble).distinct)) And you should get org.apache.spark.rdd.RDD[org.apache.spark.mllib.linalg.Vector] But if you are looking for RDD[Vector[Double]] then you can simply do val file = "/data.csv" val data = sc.textFile(file).map(line => line.split (",").map(_.toDouble).distinct.toVector)
Spark scala filter multiple rdd based on string length
I am trying to solve one of the quiz, the question is as below, Write the missing code in the given program to display the expected output to identify animals that have names with four letters. Output: Array((4,lion)) Program val a = sc.parallelize(List("dog","tiger","lion","cat","spider","eagle"),2) val b = a.keyBy(_.length) val c = sc.parallelize(List("ant","falcon","squid"),2) val d = c.keyBy(_.length) I have tried to write code in spark shell but get stuck with syntax to add 4 RDD and applying filter.
How about using the PairRDD lookup method: b.lookup(4).toArray // res1: Array[String] = Array(lion) d.lookup(4).toArray // res2: Array[String] = Array()
Preparing data for LDA in spark
I'm working on implementing a Spark LDA model (via the Scala API), and am having trouble with the necessary formatting steps for my data. My raw data (stored in a text file) is in the following format, essentially a list of tokens and the documents they correspond to. A simplified example: doc XXXXX term XXXXX 1 x 'a' x 1 x 'a' x 1 x 'b' x 2 x 'b' x 2 x 'd' x ... Where the XXXXX columns are garbage data I don't care about. I realize this is an atypical way of storing corpus data, but it's what I have. As is I hope is clear from the example, there's one line per token in the raw data (so if a given term appears 5 times in a document, that corresponds to 5 lines of text). In any case, I need to format this data as sparse term-frequency vectors for running a Spark LDA model, but am unfamiliar with Scala so having some trouble. I start with: import org.apache.spark.mllib.clustering.{LDA, DistributedLDAModel} import org.apache.spark.mllib.linalg.{Vector, Vectors} import org.apache.spark.rdd.RDD val corpus:RDD[Array[String]] = sc.textFile("path/to/data") .map(_.split('\t')).map(x => Array(x(0),x(2))) And then I get the vocabulary data I'll need to generate the sparse vectors: val vocab: RDD[String] = corpus.map(_(1)).distinct() val vocabMap: Map[String, Int] = vocab.collect().zipWithIndex.toMap What I don't know is the proper mapping function to use here such that I end up with a sparse term frequency vector for each document that I can then feed into the LDA model. I think I need something along these lines... val documents: RDD[(Long, Vector)] = corpus.groupBy(_(0)).zipWithIndex .map(x =>(x._2,Vectors.sparse(vocabMap.size, ???))) At which point I can run the actual LDA: val lda = new LDA().setK(n_topics) val ldaModel = lda.run(documents) Basically, I don't what function to apply to each group so that I can feed term frequency data (presumably as a map?) into a sparse vector. In other words, how do I fill in the ??? in the code snippet above to achieve the desired effect?
One way to handle this: make sure that spark-csv package is available load data into DataFrame and select columns of interest val df = sqlContext.read .format("com.databricks.spark.csv") .option("header", "true") .option("inferSchema", "true") // Optional, providing schema is prefered .option("delimiter", "\t") .load("foo.csv") .select($"doc".cast("long").alias("doc"), $"term") index term column: import org.apache.spark.ml.feature.StringIndexer val indexer = new StringIndexer() .setInputCol("term") .setOutputCol("termIndexed") val indexed = indexer.fit(df) .transform(df) .drop("term") .withColumn("termIndexed", $"termIndexed".cast("integer")) .groupBy($"doc", $"termIndexed") .agg(count(lit(1)).alias("cnt").cast("double")) convert to PairwiseRDD import org.apache.spark.sql.Row val pairs = indexed.map{case Row(doc: Long, term: Int, cnt: Double) => (doc, (term, cnt))} group by doc: val docs = pairs.groupByKey create feature vectors import org.apache.spark.mllib.linalg.Vectors import org.apache.spark.sql.functions.max val n = indexed.select(max($"termIndexed")).first.getInt(0) + 1 val docsWithFeatures = docs.mapValues(vs => Vectors.sparse(n, vs.toSeq)) now you have all you need to create LabeledPoints or apply additional processing
Spark DataFrame zipWithIndex
I am using a DataFrame to read in a .parquet files but than turning them into an rdd to do my normal processing I wanted to do on them. So I have my file: val dataSplit = sqlContext.parquetFile("input.parquet") val convRDD = dataSplit.rdd val columnIndex = convRDD.flatMap(r => r.zipWithIndex) I get the following error even when I convert from a dataframe to RDD: :26: error: value zipWithIndex is not a member of org.apache.spark.sql.Row Anyone know how to do what I am trying to do, essentially trying to get the value and the column index. I was thinking something like: val dataSplit = sqlContext.parquetFile(inputVal.toString) val schema = dataSplit.schema val columnIndex = dataSplit.flatMap(r => 0 until schema.length but getting stuck on the last part as not sure how to do the same of zipWithIndex.
You can simply convert Row to Seq: convRDD.flatMap(r => r.toSeq.zipWithIndex) Important thing to note here is that extracting type information becomes tricky. Row.toSeq returns Seq[Any] and resulting RDD is RDD[(Any, Int)].