Currently we are diving deeper into the Apache Spark world. We started with an implementation of a Spark-Connector written in Scala. The connector supports loading of data from ArangoDB into Spark and vice-versa. Today we release a first prototype with an aim of including our community into the development process early to build a product that fits your needs. Your feedback is more than welcome!

Usage

Setup SparkContext

First you need to initialize a SparkContext with the configuration for the Spark-Connector and the underlying Java Driver (see corresponding blog post here) to connect to your ArangoDB server.

Scala

 

Java

 

Load data from ArangoDB

To load data from ArangoDB, use the function load – from the object ArangoSpark – with the SparkContext, the name of your collection and the type of your bean to load data in. If needed, there is an additional load function with extra read options like the name of the database.

Scala

 

Java

 

Save data to ArangoDB

To save data to ArangoDB, use the function save – from the object ArangoSpark – with the SparkContext and the name of your collection. If needed, there is an additional save function with extra write options like the name of the database.

Scala / Java

 

It would be great if you try it out and give us feedback on what you think.