@@ -5,12 +5,15 @@ import com.datastax.spark.connector.cql.CassandraConnector
5
5
import org .apache .spark .sql .hive .HiveContext
6
6
import org .apache .spark .{SparkConf , SparkContext }
7
7
8
+
9
+ // For DSE it is not necessary to set connection parameters for spark.master (since it will be done
10
+ // automatically)
8
11
object WriteRead extends App {
9
12
10
13
val conf = new SparkConf ()
11
14
.setAppName(" Datastax Scala example" )
12
15
13
- CassandraConnector (conf).withSessionDo{ session =>
16
+ CassandraConnector (conf).withSessionDo { session =>
14
17
session.execute(
15
18
""" CREATE KEYSPACE IF NOT EXISTS ks WITH
16
19
| replication = {'class': 'SimpleStrategy', 'replication_factor': 1 }""" .stripMargin)
@@ -22,7 +25,7 @@ object WriteRead extends App {
22
25
val hiveContext = new HiveContext (sc)
23
26
24
27
// Write some data to C*
25
- sc.parallelize(1 to 10 ).map( x => (x,x)).saveToCassandra(" ks" , " kv" )
28
+ sc.parallelize(1 to 10 ).map(x => (x, x)).saveToCassandra(" ks" , " kv" )
26
29
27
30
// Read Data Using the Spark Context
28
31
val scReadData = sc.cassandraTable(" ks" , " kv" ).collect
@@ -31,7 +34,7 @@ object WriteRead extends App {
31
34
val sqlReadData = hiveContext
32
35
.read
33
36
.format(" org.apache.spark.sql.cassandra" )
34
- .options(Map ( " table" -> " kv" , " keyspace" -> " ks" ))
37
+ .options(Map (" table" -> " kv" , " keyspace" -> " ks" ))
35
38
.load()
36
39
.collect()
37
40
.map(row => (row.getInt(0 ), row.getInt(1 )))
0 commit comments