-
Notifications
You must be signed in to change notification settings - Fork 1
Expand file tree
/
Copy pathsession.py
More file actions
102 lines (87 loc) · 5.09 KB
/
session.py
File metadata and controls
102 lines (87 loc) · 5.09 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
from pyspark.sql import SparkSession
from .sequence import SequenceOfItems
import sys
import platform
import os
import re
import importlib.resources as pkg_resources
with pkg_resources.path("jsoniq.jars", "rumbledb-1.24.0.jar") as jar_path:
jar_path_str = "file://" + str(jar_path)
class MetaRumbleSession(type):
def __getattr__(cls, item):
if item == "builder":
return cls._builder
else:
return getattr(SparkSession, item)
class RumbleSession(object, metaclass=MetaRumbleSession):
def __init__(self, spark_session: SparkSession):
self._sparksession = spark_session
self._jrumblesession = spark_session._jvm.org.rumbledb.api.Rumble(spark_session._jsparkSession)
class Builder:
def __init__(self):
java_version = os.popen("java -version 2>&1").read()
if "version" in java_version:
match = re.search(r'version "(\d+\.\d+)', java_version)
if match:
version = match.group(1)
if not (version.startswith("17.") or version.startswith("21.")):
sys.stderr.write("**************************************************************************\n")
sys.stderr.write("[Error] RumbleDB builds on top of pyspark 4, which requires Java 17 or 21.\n")
sys.stderr.write(f"Your Java version: {version}\n")
sys.stderr.write("**************************************************************************\n")
sys.stderr.write("\n")
sys.stderr.write("What should you do?\n")
sys.stderr.write("\n")
sys.stderr.write("If you do NOT have Java 17 or 21 installed, you can download Java 17 or 21 for example from https://adoptium.net/\n")
sys.stderr.write("\n")
sys.stderr.write("Quick command for macOS: brew install --cask temurin17 or brew install --cask temurin21\n")
sys.stderr.write("Quick command for Ubuntu: apt-get install temurin-17-jdk or apt-get install temurin-21-jdk\n")
sys.stderr.write("Quick command for Windows 11: winget install EclipseAdoptium.Temurin.17.JDK or. winget install EclipseAdoptium.Temurin.21.JDK\n")
sys.stderr.write("\n")
sys.stderr.write(
"If you DO have Java 17 or 21, but the wrong version appears above, then it means you need to set your JAVA_HOME environment variable properly to point to Java 17 or 21.\n"
)
sys.stderr.write("\n")
sys.stderr.write("For macOS, try: export JAVA_HOME=$(/usr/libexec/java_home -v 17) or export JAVA_HOME=$(/usr/libexec/java_home -v 21)\n");
sys.stderr.write("\n")
sys.stderr.write("For Ubuntu, find the paths to installed versions with this command: update-alternatives --config java\n then: export JAVA_HOME=...your desired path...\n")
sys.stderr.write("\n")
sys.stderr.write("For Windows 11: look for the default Java path with 'which java' and/or look for alternate installed versions in Program Files. Then: setx /m JAVA_HOME \"...your desired path here...\"\n")
sys.exit(43)
else:
sys.stderr.write("[Error] Could not determine Java version. Please ensure Java is installed and JAVA_HOME is properly set.\n")
sys.exit(43)
self._sparkbuilder = SparkSession.builder.config("spark.jars", jar_path_str)
def getOrCreate(self):
return RumbleSession(self._sparkbuilder.getOrCreate())
def appName(self, name):
self._sparkbuilder = self._sparkbuilder.appName(name);
return self;
def master(self, url):
self._sparkbuilder = self._sparkbuilder.master(url);
return self;
def config(self, key, value):
self._sparkbuilder = self._sparkbuilder.config(key, value);
return self;
def config(self, conf):
self._sparkbuilder = self._sparkbuilder.config(conf);
return self;
def __getattr__(self, name):
res = getattr(self._sparkbuilder, name);
return res;
_builder = Builder()
def bindDataFrameAsVariable(self, name: str, df):
conf = self._jrumblesession.getConfiguration();
if not name.startswith("$"):
raise ValueError("Variable name must start with a dollar symbol ('$').")
name = name[1:]
if(hasattr(df, "_get_object_id")):
conf.setExternalVariableValue(name, df);
else:
conf.setExternalVariableValue(name, df._jdf);
return self;
def jsoniq(self, str):
sequence = self._jrumblesession.runQuery(str);
return SequenceOfItems(sequence, self._sparksession.sparkContext);
def __getattr__(self, item):
return getattr(self._sparksession, item)