accumulate all values for a given key in spark
For a given key collect all the values, which can be later use for applying some custom logic like ( average , max , min, top n, expression evaluation) in Spark
For a given key collect all the values, which can be later use for applying some custom logic like ( average , max , min, top n, expression evaluation) in Spark
Read from Aerospike with a spark application via mapPartitions in Spark
Read From Aerospike Using Spark via Map Transformation in Spark
Write to Aerospike from spark via MapPartitions
Read from Hdfs and write to Aerospike from Spark via Map Transformation
For integration test a embedded zookeeper module
Aerospike Unit Test framework to run test on Aerospike as part of maven module
Avro to Row conversion
Understanding Spark Serialization , and in the process try to understand when to use lambada function , static,anonymous class and transient references
Spark Map Vs mapPartitions