diff --git a/.gitlab-ci.yml b/.gitlab-ci.yml index 30145d863d56d74c735b5f90cbbfb41bd7b1bcd9..8d5fabc73c3790773f62181011adbc219ee716b3 100644 --- a/.gitlab-ci.yml +++ b/.gitlab-ci.yml @@ -34,14 +34,10 @@ test: variables: PIP_CACHE_DIR: "$CI_PROJECT_DIR/.cache/pip" - ARKINDEX_API_SCHEMA_URL: schema.yml before_script: - pip install tox - # Download OpenAPI schema from last backend build - - curl https://assets.teklia.com/arkindex/openapi.yml > schema.yml - # Add system deps for opencv - apt-get update -q - apt-get install -q -y libgl1 diff --git a/docs/get_started/training.md b/docs/get_started/training.md index b996f018cee61e77f086a41106f4784364a0bece..6684fd05bc603aae9791a7ed5a96f1278d277157 100644 --- a/docs/get_started/training.md +++ b/docs/get_started/training.md @@ -6,7 +6,7 @@ There are a several steps to follow when training a DAN model. The data must be extracted and formatted for training. To extract the data, DAN uses an Arkindex export database in SQLite format. You will need to: -1. Structure the data into folders (`train` / `val` / `test`) in [Arkindex](https://arkindex.teklia.com/). +1. Structure the data into folders (`train` / `val` / `test`) in [Arkindex](https://demo.arkindex.org/). 1. [Export the project](https://doc.arkindex.org/howto/export/) in SQLite format. 1. Extract the data with the [extract command](../usage/datasets/extract.md). 1. Format the data with the [format command](../usage/datasets/format.md). diff --git a/docs/usage/datasets/index.md b/docs/usage/datasets/index.md index a856f6c099d9620d01f38d0cc72c1cafc460b4e3..8e6a21cc2b878a2cd27528b8d986993777b39756 100644 --- a/docs/usage/datasets/index.md +++ b/docs/usage/datasets/index.md @@ -3,7 +3,7 @@ Two operations are available through subcommands: `teklia-dan dataset extract` -: To extract a dataset from Arkindex using its [Python API](https://arkindex.teklia.com/api-docs/). More details in [the dedicated section](./extract.md). +: To extract a dataset from Arkindex using its [Python API](https://demo.arkindex.org/api-docs/). More details in [the dedicated section](./extract.md). `teklia-dan dataset format` : To format datasets for training. More details in [the dedicated section](./format.md). diff --git a/tests/conftest.py b/tests/conftest.py index f65fcb74dd61b3d8aa1e831f997fdaeec652409d..c73aac32ae13e4da334062b76b52eb3e724db39f 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -1,5 +1,4 @@ # -*- coding: utf-8 -*- -import os from pathlib import Path import pytest @@ -14,22 +13,6 @@ from dan.transforms import Preprocessing FIXTURES = Path(__file__).resolve().parent / "data" -@pytest.fixture(autouse=True) -def setup_environment(responses): - """Setup needed environment variables""" - - # Allow accessing remote API schemas - # defaulting to the prod environment - schema_url = os.environ.get( - "ARKINDEX_API_SCHEMA_URL", - "https://arkindex.teklia.com/api/v1/openapi/?format=openapi-json", - ) - responses.add_passthru(schema_url) - - # Set schema url in environment - os.environ["ARKINDEX_API_SCHEMA_URL"] = schema_url - - @pytest.fixture def database_path(): return FIXTURES / "export.sqlite" diff --git a/tox.ini b/tox.ini index aedc8cf03e4c8e19e4a43fcf6c2a3061a2de5db5..0b48a6dd22ee59fd5aa372e438ad305934837a38 100644 --- a/tox.ini +++ b/tox.ini @@ -2,11 +2,9 @@ envlist = teklia-dan [testenv] -passenv = ARKINDEX_API_SCHEMA_URL commands = pytest {posargs} deps = pytest - pytest-responses -rrequirements.txt