214 lines
6.1 KiB
Python
214 lines
6.1 KiB
Python
import os
|
|
import shlex
|
|
import subprocess
|
|
import time
|
|
|
|
import pytest
|
|
|
|
from pandas.compat import (
|
|
is_ci_environment,
|
|
is_platform_arm,
|
|
is_platform_mac,
|
|
is_platform_windows,
|
|
)
|
|
import pandas.util._test_decorators as td
|
|
|
|
import pandas._testing as tm
|
|
|
|
from pandas.io.parsers import read_csv
|
|
|
|
|
|
@pytest.fixture
|
|
def tips_file(datapath):
|
|
"""Path to the tips dataset"""
|
|
return datapath("io", "data", "csv", "tips.csv")
|
|
|
|
|
|
@pytest.fixture
|
|
def jsonl_file(datapath):
|
|
"""Path to a JSONL dataset"""
|
|
return datapath("io", "parser", "data", "items.jsonl")
|
|
|
|
|
|
@pytest.fixture
|
|
def salaries_table(datapath):
|
|
"""DataFrame with the salaries dataset"""
|
|
return read_csv(datapath("io", "parser", "data", "salaries.csv"), sep="\t")
|
|
|
|
|
|
@pytest.fixture
|
|
def feather_file(datapath):
|
|
return datapath("io", "data", "feather", "feather-0_3_1.feather")
|
|
|
|
|
|
@pytest.fixture
|
|
def s3so(worker_id):
|
|
if is_ci_environment():
|
|
url = "http://localhost:5000/"
|
|
else:
|
|
worker_id = "5" if worker_id == "master" else worker_id.lstrip("gw")
|
|
url = f"http://127.0.0.1:555{worker_id}/"
|
|
return {"client_kwargs": {"endpoint_url": url}}
|
|
|
|
|
|
@pytest.fixture(scope="session")
|
|
def s3_base(worker_id):
|
|
"""
|
|
Fixture for mocking S3 interaction.
|
|
|
|
Sets up moto server in separate process locally
|
|
Return url for motoserver/moto CI service
|
|
"""
|
|
pytest.importorskip("s3fs")
|
|
pytest.importorskip("boto3")
|
|
|
|
with tm.ensure_safe_environment_variables():
|
|
# temporary workaround as moto fails for botocore >= 1.11 otherwise,
|
|
# see https://github.com/spulec/moto/issues/1924 & 1952
|
|
os.environ.setdefault("AWS_ACCESS_KEY_ID", "foobar_key")
|
|
os.environ.setdefault("AWS_SECRET_ACCESS_KEY", "foobar_secret")
|
|
if is_ci_environment():
|
|
if is_platform_arm() or is_platform_mac() or is_platform_windows():
|
|
# NOT RUN on Windows/macOS/ARM, only Ubuntu
|
|
# - subprocess in CI can cause timeouts
|
|
# - GitHub Actions do not support
|
|
# container services for the above OSs
|
|
# - CircleCI will probably hit the Docker rate pull limit
|
|
pytest.skip(
|
|
"S3 tests do not have a corresponding service in "
|
|
"Windows, macOS or ARM platforms"
|
|
)
|
|
else:
|
|
yield "http://localhost:5000"
|
|
else:
|
|
requests = pytest.importorskip("requests")
|
|
pytest.importorskip("moto", minversion="1.3.14")
|
|
pytest.importorskip("flask") # server mode needs flask too
|
|
|
|
# Launching moto in server mode, i.e., as a separate process
|
|
# with an S3 endpoint on localhost
|
|
|
|
worker_id = "5" if worker_id == "master" else worker_id.lstrip("gw")
|
|
endpoint_port = f"555{worker_id}"
|
|
endpoint_uri = f"http://127.0.0.1:{endpoint_port}/"
|
|
|
|
# pipe to null to avoid logging in terminal
|
|
with subprocess.Popen(
|
|
shlex.split(f"moto_server s3 -p {endpoint_port}"),
|
|
stdout=subprocess.DEVNULL,
|
|
stderr=subprocess.DEVNULL,
|
|
) as proc:
|
|
timeout = 5
|
|
while timeout > 0:
|
|
try:
|
|
# OK to go once server is accepting connections
|
|
r = requests.get(endpoint_uri)
|
|
if r.ok:
|
|
break
|
|
except Exception:
|
|
pass
|
|
timeout -= 0.1
|
|
time.sleep(0.1)
|
|
yield endpoint_uri
|
|
|
|
proc.terminate()
|
|
|
|
|
|
@pytest.fixture
|
|
def s3_resource(s3_base, tips_file, jsonl_file, feather_file):
|
|
"""
|
|
Sets up S3 bucket with contents
|
|
|
|
The primary bucket name is "pandas-test". The following datasets
|
|
are loaded.
|
|
|
|
- tips.csv
|
|
- tips.csv.gz
|
|
- tips.csv.bz2
|
|
- items.jsonl
|
|
|
|
A private bucket "cant_get_it" is also created. The boto3 s3 resource
|
|
is yielded by the fixture.
|
|
"""
|
|
import boto3
|
|
import s3fs
|
|
|
|
test_s3_files = [
|
|
("tips#1.csv", tips_file),
|
|
("tips.csv", tips_file),
|
|
("tips.csv.gz", tips_file + ".gz"),
|
|
("tips.csv.bz2", tips_file + ".bz2"),
|
|
("items.jsonl", jsonl_file),
|
|
("simple_dataset.feather", feather_file),
|
|
]
|
|
|
|
def add_tips_files(bucket_name):
|
|
for s3_key, file_name in test_s3_files:
|
|
with open(file_name, "rb") as f:
|
|
cli.put_object(Bucket=bucket_name, Key=s3_key, Body=f)
|
|
|
|
bucket = "pandas-test"
|
|
conn = boto3.resource("s3", endpoint_url=s3_base)
|
|
cli = boto3.client("s3", endpoint_url=s3_base)
|
|
|
|
try:
|
|
cli.create_bucket(Bucket=bucket)
|
|
except Exception:
|
|
# OK is bucket already exists
|
|
pass
|
|
try:
|
|
cli.create_bucket(Bucket="cant_get_it", ACL="private")
|
|
except Exception:
|
|
# OK is bucket already exists
|
|
pass
|
|
timeout = 2
|
|
while not cli.list_buckets()["Buckets"] and timeout > 0:
|
|
time.sleep(0.1)
|
|
timeout -= 0.1
|
|
|
|
add_tips_files(bucket)
|
|
add_tips_files("cant_get_it")
|
|
s3fs.S3FileSystem.clear_instance_cache()
|
|
yield conn
|
|
|
|
s3 = s3fs.S3FileSystem(client_kwargs={"endpoint_url": s3_base})
|
|
|
|
try:
|
|
s3.rm(bucket, recursive=True)
|
|
except Exception:
|
|
pass
|
|
try:
|
|
s3.rm("cant_get_it", recursive=True)
|
|
except Exception:
|
|
pass
|
|
timeout = 2
|
|
while cli.list_buckets()["Buckets"] and timeout > 0:
|
|
time.sleep(0.1)
|
|
timeout -= 0.1
|
|
|
|
|
|
_compression_formats_params = [
|
|
(".no_compress", None),
|
|
("", None),
|
|
(".gz", "gzip"),
|
|
(".GZ", "gzip"),
|
|
(".bz2", "bz2"),
|
|
(".BZ2", "bz2"),
|
|
(".zip", "zip"),
|
|
(".ZIP", "zip"),
|
|
(".xz", "xz"),
|
|
(".XZ", "xz"),
|
|
pytest.param((".zst", "zstd"), marks=td.skip_if_no("zstandard")),
|
|
pytest.param((".ZST", "zstd"), marks=td.skip_if_no("zstandard")),
|
|
]
|
|
|
|
|
|
@pytest.fixture(params=_compression_formats_params[1:])
|
|
def compression_format(request):
|
|
return request.param
|
|
|
|
|
|
@pytest.fixture(params=_compression_formats_params)
|
|
def compression_ext(request):
|
|
return request.param[0]
|