# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. """ Utility functions for testing """ import contextlib import decimal import gc import os import random import re import shutil import signal import socket import string import subprocess import sys import time import pytest import pyarrow as pa import pyarrow.fs def randsign(): """Randomly choose either 1 or -1. Returns ------- sign : int """ return random.choice((-1, 1)) @contextlib.contextmanager def random_seed(seed): """Set the random seed inside of a context manager. Parameters ---------- seed : int The seed to set Notes ----- This function is useful when you want to set a random seed but not affect the random state of other functions using the random module. """ original_state = random.getstate() random.seed(seed) try: yield finally: random.setstate(original_state) def randdecimal(precision, scale): """Generate a random decimal value with specified precision and scale. Parameters ---------- precision : int The maximum number of digits to generate. Must be an integer between 1 and 38 inclusive. scale : int The maximum number of digits following the decimal point. Must be an integer greater than or equal to 0. Returns ------- decimal_value : decimal.Decimal A random decimal.Decimal object with the specified precision and scale. """ assert 1 <= precision <= 38, 'precision must be between 1 and 38 inclusive' if scale < 0: raise ValueError( 'randdecimal does not yet support generating decimals with ' 'negative scale' ) max_whole_value = 10 ** (precision - scale) - 1 whole = random.randint(-max_whole_value, max_whole_value) if not scale: return decimal.Decimal(whole) max_fractional_value = 10 ** scale - 1 fractional = random.randint(0, max_fractional_value) return decimal.Decimal( '{}.{}'.format(whole, str(fractional).rjust(scale, '0')) ) def random_ascii(length): return bytes([random.randint(65, 122) for i in range(length)]) def rands(nchars): """ Generate one random string. """ RANDS_CHARS = list(string.ascii_letters + string.digits) return "".join(random.choice(RANDS_CHARS) for i in range(nchars)) def memory_leak_check(f, metric='rss', threshold=1 << 17, iterations=10, check_interval=1): """ Execute the function and try to detect a clear memory leak either internal to Arrow or caused by a reference counting problem in the Python binding implementation. Raises exception if a leak detected Parameters ---------- f : callable Function to invoke on each iteration metric : {'rss', 'vms', 'shared'}, default 'rss' Attribute of psutil.Process.memory_info to use for determining current memory use threshold : int, default 128K Threshold in number of bytes to consider a leak iterations : int, default 10 Total number of invocations of f check_interval : int, default 1 Number of invocations of f in between each memory use check """ import psutil proc = psutil.Process() def _get_use(): gc.collect() return getattr(proc.memory_info(), metric) baseline_use = _get_use() def _leak_check(): current_use = _get_use() if current_use - baseline_use > threshold: raise Exception("Memory leak detected. " "Departure from baseline {} after {} iterations" .format(current_use - baseline_use, i)) for i in range(iterations): f() if i % check_interval == 0: _leak_check() def get_modified_env_with_pythonpath(): # Prepend pyarrow root directory to PYTHONPATH env = os.environ.copy() existing_pythonpath = env.get('PYTHONPATH', '') module_path = os.path.abspath( os.path.dirname(os.path.dirname(pa.__file__))) if existing_pythonpath: new_pythonpath = os.pathsep.join((module_path, existing_pythonpath)) else: new_pythonpath = module_path env['PYTHONPATH'] = new_pythonpath return env def invoke_script(script_name, *args): subprocess_env = get_modified_env_with_pythonpath() dir_path = os.path.dirname(os.path.realpath(__file__)) python_file = os.path.join(dir_path, script_name) cmd = [sys.executable, python_file] cmd.extend(args) subprocess.check_call(cmd, env=subprocess_env) @contextlib.contextmanager def changed_environ(name, value): """ Temporarily set environment variable *name* to *value*. """ orig_value = os.environ.get(name) os.environ[name] = value try: yield finally: if orig_value is None: del os.environ[name] else: os.environ[name] = orig_value @contextlib.contextmanager def change_cwd(path): curdir = os.getcwd() os.chdir(str(path)) try: yield finally: os.chdir(curdir) @contextlib.contextmanager def disabled_gc(): gc.disable() try: yield finally: gc.enable() def _filesystem_uri(path): # URIs on Windows must follow 'file:///C:...' or 'file:/C:...' patterns. if os.name == 'nt': uri = 'file:///{}'.format(path) else: uri = 'file://{}'.format(path) return uri class FSProtocolClass: def __init__(self, path): self._path = path def __fspath__(self): return str(self._path) class ProxyHandler(pyarrow.fs.FileSystemHandler): """ A dataset handler that proxies to an underlying filesystem. Useful to partially wrap an existing filesystem with partial changes. """ def __init__(self, fs): self._fs = fs def __eq__(self, other): if isinstance(other, ProxyHandler): return self._fs == other._fs return NotImplemented def __ne__(self, other): if isinstance(other, ProxyHandler): return self._fs != other._fs return NotImplemented def get_type_name(self): return "proxy::" + self._fs.type_name def normalize_path(self, path): return self._fs.normalize_path(path) def get_file_info(self, paths): return self._fs.get_file_info(paths) def get_file_info_selector(self, selector): return self._fs.get_file_info(selector) def create_dir(self, path, recursive): return self._fs.create_dir(path, recursive=recursive) def delete_dir(self, path): return self._fs.delete_dir(path) def delete_dir_contents(self, path, missing_dir_ok): return self._fs.delete_dir_contents(path, missing_dir_ok=missing_dir_ok) def delete_root_dir_contents(self): return self._fs.delete_dir_contents("", accept_root_dir=True) def delete_file(self, path): return self._fs.delete_file(path) def move(self, src, dest): return self._fs.move(src, dest) def copy_file(self, src, dest): return self._fs.copy_file(src, dest) def open_input_stream(self, path): return self._fs.open_input_stream(path) def open_input_file(self, path): return self._fs.open_input_file(path) def open_output_stream(self, path, metadata): return self._fs.open_output_stream(path, metadata=metadata) def open_append_stream(self, path, metadata): return self._fs.open_append_stream(path, metadata=metadata) @contextlib.contextmanager def signal_wakeup_fd(*, warn_on_full_buffer=False): # Use a socket pair, rather a self-pipe, so that select() can be used # on Windows. r, w = socket.socketpair() old_fd = None try: r.setblocking(False) w.setblocking(False) old_fd = signal.set_wakeup_fd( w.fileno(), warn_on_full_buffer=warn_on_full_buffer) yield r finally: if old_fd is not None: signal.set_wakeup_fd(old_fd) r.close() w.close() def _ensure_minio_component_version(component, minimum_year): full_args = [component, '--version'] with subprocess.Popen(full_args, stdout=subprocess.PIPE, stderr=subprocess.PIPE, encoding='utf-8') as proc: if proc.wait(10) != 0: return False stdout = proc.stdout.read() pattern = component + r' version RELEASE\.(\d+)-.*' version_match = re.search(pattern, stdout) if version_match: version_year = version_match.group(1) return int(version_year) >= minimum_year else: raise FileNotFoundError( "minio component older than the minimum year") def _wait_for_minio_startup(mcdir, address, access_key, secret_key): start = time.time() while time.time() - start < 10: try: _run_mc_command(mcdir, 'alias', 'set', 'myminio', f'http://{address}', access_key, secret_key) return except ChildProcessError: time.sleep(1) raise Exception("mc command could not connect to local minio") def _run_mc_command(mcdir, *args): full_args = ['mc', '-C', mcdir] + list(args) with subprocess.Popen(full_args, stdout=subprocess.PIPE, stderr=subprocess.PIPE, encoding='utf-8') as proc: retval = proc.wait(10) cmd_str = ' '.join(full_args) print(f'Cmd: {cmd_str}') print(f' Return: {retval}') print(f' Stdout: {proc.stdout.read()}') print(f' Stderr: {proc.stderr.read()}') if retval != 0: raise ChildProcessError("Could not run mc") def _configure_s3_limited_user(s3_server, policy, username, password): """ Attempts to use the mc command to configure the minio server with a special user limited:limited123 which does not have permission to create buckets. This mirrors some real life S3 configurations where users are given strict permissions. Arrow S3 operations should still work in such a configuration (e.g. see ARROW-13685) """ if sys.platform == 'win32': # Can't rely on FileNotFound check because # there is sometimes an mc command on Windows # which is unrelated to the minio mc pytest.skip('The mc command is not installed on Windows') try: # ensuring version of mc and minio for the capabilities we need _ensure_minio_component_version('mc', 2021) _ensure_minio_component_version('minio', 2021) tempdir = s3_server['tempdir'] host, port, access_key, secret_key = s3_server['connection'] address = '{}:{}'.format(host, port) mcdir = os.path.join(tempdir, 'mc') if os.path.exists(mcdir): shutil.rmtree(mcdir) os.mkdir(mcdir) policy_path = os.path.join(tempdir, 'limited-buckets-policy.json') with open(policy_path, mode='w') as policy_file: policy_file.write(policy) # The s3_server fixture starts the minio process but # it takes a few moments for the process to become available _wait_for_minio_startup(mcdir, address, access_key, secret_key) # Create a limited user with a specific policy ... _run_mc_command(mcdir, 'admin', 'user', 'add', 'myminio/', username, password) _run_mc_command(mcdir, 'admin', 'policy', 'create', 'myminio/', 'no-create-buckets', policy_path) _run_mc_command(mcdir, 'admin', 'policy', 'attach', 'myminio/', 'no-create-buckets', '--user', username) # ... and a sample bucket for that user to write to _run_mc_command(mcdir, 'mb', 'myminio/existing-bucket', '--ignore-existing') except FileNotFoundError: pytest.skip("Configuring limited s3 user failed") def windows_has_tzdata(): """ This is the default location where tz.cpp will look for (until we make this configurable at run-time) """ tzdata_bool = False if "PYARROW_TZDATA_PATH" in os.environ: tzdata_bool = os.path.exists(os.environ['PYARROW_TZDATA_PATH']) if not tzdata_bool: tzdata_path = os.path.expandvars(r"%USERPROFILE%\Downloads\tzdata") tzdata_bool = os.path.exists(tzdata_path) return tzdata_bool