From 596ac8930efcfcffd0857d3e88b2e3a9e9644980 Mon Sep 17 00:00:00 2001 From: nthanvi Date: Tue, 17 May 2016 14:43:35 +0530 Subject: [PATCH] [SNAPPYDATA] Use SnappyContext as default SQLContext on shell (#35) Conflicts: python/pyspark/shell.py --- python/pyspark/shell.py | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/python/pyspark/shell.py b/python/pyspark/shell.py index b5fcf7092d93a..516185f10cdce 100644 --- a/python/pyspark/shell.py +++ b/python/pyspark/shell.py @@ -31,6 +31,8 @@ from pyspark import SparkConf from pyspark.context import SparkContext from pyspark.sql import SparkSession, SQLContext +from pyspark.sql.snappy import SnappyContext +from pyspark.storagelevel import StorageLevel if os.environ.get("SPARK_EXECUTOR_URI"): SparkContext.setSystemProperty("spark.executor.uri", os.environ["SPARK_EXECUTOR_URI"]) @@ -38,6 +40,8 @@ SparkContext._ensure_initialized() try: + sqlContext = SnappyContext(sc) +except py4j.protocol.Py4JError: # Try to access HiveConf, it will raise exception if Hive is not added conf = SparkConf() if conf.get('spark.sql.catalogImplementation', 'hive').lower() == 'hive':