Connect Apache Spark SQL to Node.js on Linux via JDBC Driver

Connect Apache Spark SQL to Node.js on Linux via JDBC Driver

Posted on August 02, 2016 0 Comments

This tutorial explains how to access Apache Spark SQL data from a Node.js application using DataDirect Apache Spark SQL JDBC driver on a Linux machine/server.

Apache Spark is changing the way Big Data is accessed and processed. While MapReduce was a good implementation for processing and generating large data sets with a parallel, distributed algorithm on a cluster, it was not optimized for interactive data analysis that involves iterative algorithms. Spark was designed to overcome this shortcoming.

As you implement Apache Spark in your organization, we understand that you need ways to connect your Apache Spark to other JDBC applications. Apache Spark SQL allows you to connect with any JDBC data source. We put together a tutorial that explains how you can connect to a Node.js application on Linux using a Spark SQL JDBC driver.

If you are looking to connect to a Node.js ODBC application using a Spark SQL ODBC driver, visit this tutorial.

Before You Start

  1. Make sure that you have Java installed on your machine. You can check this by running the command java -version on your terminal.
  2. If Java is installed, the current version of Java will be displayed. If not, please install Java before proceeding to next steps.
  3. Make sure you have Apache Spark SQL installed on your machine. You can download and install Apache Spark SQL pre-built for Hadoop here.

Install DataDirect Spark SQL JDBC Driver

  1. Download the DataDirect Spark SQL JDBC driver from here.
  2. Extract the contents from the downloaded package by opening a terminal at the downloaded location and running the following command:
  3. unzip
  4. To install the driver, you have to execute the .jar package by running the following command in terminal:

  5. This will launch an interactive Java installer, which you can use to install the Spark SQL JDBC driver to your desired location as either a licensed or evaluation installation.

Load Data in to Spark SQL

  1. For the purpose of the tutorial, I will be loading the data from a CSV file that can be found here.
  2. Start the Spark shell using the following command, which has been configured to run the Thrift server in single-session mode as I am only going to register the imported data as Temporary table. I am also including a package that can be used to import data from the CSV, as it is not supported natively:

    spark-shell --conf spark.sql.hive.thriftServer.singleSession=true --packages com.databricks:spark-csv_2.11:1.4.0

  3. Once the Spark shell starts successfully, run the following commands to import the data from the CSV and register it as temporary table:

import org.apache.spark.sql._
import org.apache.spark.sql.hive._
import org.apache.spark.sql.hive.thriftserver._
//Read from CSV
val df ="com.databricks.spark.csv").option("inferSchema","true").option("header","true").load("/path/to/InsuranceData.csv")
//Check if CSV was imported successfully
//Register Temp Table
sqlContext.sql("select count(*) from InsuranceData").show()
val hc = sqlContext.asInstanceOf[HiveContext]


Connect to Your Data from Node.js

  1. In your Node.js application, install the module node-jdbc using npm. Read this page on how to install node-jdbc module.
  2. Copy the SparkSQL JDBC driver from /install_dir/Progress/JDBC_XX/lib to your project library.
  3. You can now access the data from Spark SQL using DataDirect Spark SQL JDBC driver by loading the JDBC module in your code. The following code snippet demonstrates on how you can do it:

    var JDBC = require('jdbc');
    var jinst = require('jdbc/lib/jinst');
    var asyncjs = require('async');
    if (!jinst.isJvmCreated()) {
      var config = {
        // SparkSQL configuration to your server
        url: 'jdbc:datadirect:sparksql://<;hostname>:<port>;DatabaseName=default',
        drivername: 'com.ddtek.jdbc.sparksql.SparkSQLDriver',
        minpoolsize: 1,
        maxpoolsize: 100,
        user: 'username',
        password: 'password',
        properties: {}
      var sparksqldb = new JDBC(config);
      sparksqldb.initialize(function(err) {
        if (err) {
      sparksqldb.reserve(function(err, connObj) {
        if (connObj) {
          console.log("Using connection: " + connObj.uuid);
          var conn = connObj.conn;
          // Query the database.
            function(callback) {
              // Select statement example.
              conn.createStatement(function(err, statement) {
                if (err) {
                } else {
                  statement.setFetchSize(100, function(err) {
                    if (err) {
                    } else {
                //Execute a query
                      statement.executeQuery("SELECT * FROM InsuranceData;",
                          function(err, resultset) {
                            if (err) {
                            } else {
                              resultset.toObjArray(function(err, results) {
                                //Printing number of records
                                if (results.length > 0) {
                                  console.log("Record count: " + results.length);
                                callback(null, resultset);
          ], function(err, results) {
            // Results can also be processed here.
            // Release the connection back to the pool.
            sparksqldb.release(connObj, function(err) {
              if (err) {
  4. Notice the method setupClasspath, where you would have to give a path to the DataDirect SparkSQL JDBC driver. When you run the above code, it should print the count of records in your temporary table to console.

Next Steps

The DataDirect SparkSQL JDBC driver is the best-in-class certified connectivity solution for Apache Spark SQL. For more information about the driver, visit our Spark SQL JDBC driver page and try it free for 15 days. You can also learn more about our other solutions for Apache and other big data frameworks here. Subscribe to our blog via email or RSS feed for more tutorials.

Try Now


Saikrishna Teja Bobba

Saikrishna is a DataDirect Developer Evangelist at Progress. Prior to working at Progress, he worked as Software Engineer for 3 years after getting his undergraduate degree, and recently graduated from NC State University with Masters in Computer Science. His interests are in the areas of Data Connectivity, SaaS and Mobile App Development.


Comments are disabled in preview mode.

Sitefinity Training and Certification Now Available.

Let our experts teach you how to use Sitefinity's best-in-class features to deliver compelling digital experiences.

Learn More
Latest Stories
in Your Inbox

Subscribe to get all the news, info and tutorials you need to build better business apps and sites

Loading animation