forked from datafold/data-diff
-
Notifications
You must be signed in to change notification settings - Fork 1
Expand file tree
/
Copy pathcommon.py
More file actions
182 lines (137 loc) · 6.31 KB
/
common.py
File metadata and controls
182 lines (137 loc) · 6.31 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
import hashlib
import logging
import os
import random
import re
import string
import subprocess
import unittest
from collections.abc import Callable
from parameterized import parameterized_class
from data_diff import connect
from data_diff import databases as db
from data_diff.databases.base import Database
from data_diff.queries.api import table
from data_diff.query_utils import drop_table
from data_diff.table_segment import TableSegment
# We write 'or None' because Github sometimes creates empty env vars for secrets
TEST_MYSQL_CONN_STRING: str = os.environ.get("DATADIFF_MYSQL_URI") or "mysql://mysql:Password1@localhost/mysql"
TEST_POSTGRESQL_CONN_STRING: str = (
os.environ.get("DATADIFF_POSTGRESQL_URI") or "postgresql://postgres:Password1@localhost/postgres"
)
TEST_SNOWFLAKE_CONN_STRING: str = os.environ.get("DATADIFF_SNOWFLAKE_URI") or None
TEST_PRESTO_CONN_STRING: str = os.environ.get("DATADIFF_PRESTO_URI") or None
TEST_BIGQUERY_CONN_STRING: str = os.environ.get("DATADIFF_BIGQUERY_URI") or None
TEST_REDSHIFT_CONN_STRING: str = os.environ.get("DATADIFF_REDSHIFT_URI") or None
TEST_ORACLE_CONN_STRING: str = None
TEST_DATABRICKS_CONN_STRING: str = os.environ.get("DATADIFF_DATABRICKS_URI") or None
TEST_TRINO_CONN_STRING: str = os.environ.get("DATADIFF_TRINO_URI") or None
TEST_CLICKHOUSE_CONN_STRING: str = os.environ.get("DATADIFF_CLICKHOUSE_URI") or None
TEST_VERTICA_CONN_STRING: str = os.environ.get("DATADIFF_VERTICA_URI") or None
TEST_DUCKDB_CONN_STRING: str = "duckdb://main:@:memory:"
TEST_MSSQL_CONN_STRING: str = os.environ.get("DATADIFF_MSSQL_URI") or None
DEFAULT_N_SAMPLES = 50
N_SAMPLES = int(os.environ.get("N_SAMPLES", DEFAULT_N_SAMPLES))
BENCHMARK = os.environ.get("BENCHMARK", False)
N_THREADS = int(os.environ.get("N_THREADS", 1))
TEST_ACROSS_ALL_DBS = os.environ.get("TEST_ACROSS_ALL_DBS", True) # Should we run the full db<->db test suite?
def get_git_revision_short_hash() -> str:
return subprocess.check_output(["git", "rev-parse", "--short", "HEAD"]).decode("ascii").strip()
GIT_REVISION = get_git_revision_short_hash()
level = logging.ERROR
if os.environ.get("LOG_LEVEL", False):
level = getattr(logging, os.environ["LOG_LEVEL"].upper())
logging.basicConfig(level=level)
logging.getLogger("hashdiff_tables").setLevel(level)
logging.getLogger("joindiff_tables").setLevel(level)
logging.getLogger("diff_tables").setLevel(level)
logging.getLogger("table_segment").setLevel(level)
logging.getLogger("database").setLevel(level)
try:
from tests.local_settings import *
except ImportError:
pass # No local settings
CONN_STRINGS = {
db.BigQuery: TEST_BIGQUERY_CONN_STRING,
db.MySQL: TEST_MYSQL_CONN_STRING,
db.PostgreSQL: TEST_POSTGRESQL_CONN_STRING,
db.Snowflake: TEST_SNOWFLAKE_CONN_STRING,
db.Redshift: TEST_REDSHIFT_CONN_STRING,
db.Oracle: TEST_ORACLE_CONN_STRING,
db.Presto: TEST_PRESTO_CONN_STRING,
db.Databricks: TEST_DATABRICKS_CONN_STRING,
db.Trino: TEST_TRINO_CONN_STRING,
db.Clickhouse: TEST_CLICKHOUSE_CONN_STRING,
db.Vertica: TEST_VERTICA_CONN_STRING,
db.DuckDB: TEST_DUCKDB_CONN_STRING,
db.MsSQL: TEST_MSSQL_CONN_STRING,
}
def get_conn(cls: type) -> Database:
return connect(CONN_STRINGS[cls], N_THREADS)
def _print_used_dbs():
used = {k.__name__ for k, v in CONN_STRINGS.items() if v is not None}
unused = {k.__name__ for k, v in CONN_STRINGS.items() if v is None}
print(f"Testing databases: {', '.join(used)}")
if unused:
logging.info(f"Connection not configured; skipping tests for: {', '.join(unused)}")
if TEST_ACROSS_ALL_DBS:
logging.info(
f"Full tests enabled (every db<->db). May take very long when many dbs are involved. ={TEST_ACROSS_ALL_DBS}"
)
_print_used_dbs()
CONN_STRINGS = {k: v for k, v in CONN_STRINGS.items() if v is not None}
def random_table_suffix() -> str:
char_set = string.ascii_lowercase + string.digits
suffix = "_"
suffix += "".join(random.choice(char_set) for _ in range(5))
return suffix
def str_to_checksum(str: str):
# hello world
# => 5eb63bbbe01eeed093cb22bb8f5acdc3
# => cb22bb8f5acdc3
# => 273350391345368515 - offset (see db.CHECKSUM_OFFSET)
m = hashlib.md5()
m.update(str.encode("utf-8")) # encode to binary
md5 = m.hexdigest()
# 0-indexed, unlike DBs which are 1-indexed here, so +1 in dbs
half_pos = db.MD5_HEXDIGITS - db.CHECKSUM_HEXDIGITS
return int(md5[half_pos:], 16) - db.CHECKSUM_OFFSET
class DiffTestCase(unittest.TestCase):
"""Sets up two tables for diffing"""
db_cls = None
src_schema = None
dst_schema = None
def setUp(self):
assert self.db_cls, self.db_cls
self.connection = get_conn(self.db_cls)
table_suffix = random_table_suffix()
self.table_src_name = f"src{table_suffix}"
self.table_dst_name = f"dst{table_suffix}"
self.table_src_path = self.connection.dialect.parse_table_name(self.table_src_name)
self.table_dst_path = self.connection.dialect.parse_table_name(self.table_dst_name)
drop_table(self.connection, self.table_src_path)
drop_table(self.connection, self.table_dst_path)
self.src_table = table(self.table_src_path, schema=self.src_schema)
self.dst_table = table(self.table_dst_path, schema=self.dst_schema)
if self.src_schema:
self.connection.query(self.src_table.create())
if self.dst_schema:
self.connection.query(self.dst_table.create())
return super().setUp()
def tearDown(self):
drop_table(self.connection, self.table_src_path)
drop_table(self.connection, self.table_dst_path)
def _parameterized_class_per_conn(test_databases):
test_databases = set(test_databases)
names = [(cls.__name__, cls) for cls in CONN_STRINGS if cls in test_databases]
return parameterized_class(("name", "db_cls"), names)
def apply_to_each_database(databases) -> Callable:
def _test_per_database(cls):
return _parameterized_class_per_conn(databases)(cls)
return _test_per_database
def table_segment(database, table_path, key_columns, *args, **kw):
if isinstance(key_columns, str):
key_columns = (key_columns,)
return TableSegment(database, table_path, key_columns, *args, **kw)
def ansi_stdout_cleanup(ansi_input) -> str:
return re.sub(r"\x1B\[[0-?]*[ -/]*[@-~]", "", ansi_input)