@@ -5,13 +5,6 @@ package:
5
5
rm -Rf dist; \
6
6
python -m build
7
7
8
- pypi :
9
- python -m pip install twine -U
10
- python -m twine upload $(plugin ) /dist/*
11
-
12
- install : package
13
- cd $(plugin ) && pip install -U dist/* .whl
14
-
15
8
install-pip-setuptools :
16
9
python -m pip install -U pip setuptools wheel
17
10
@@ -25,46 +18,14 @@ mypy:
25
18
test :
26
19
cd $(plugin ) && pytest tests --cov-config pyproject.toml --numprocesses 4 --dist loadfile
27
20
28
- # Run test_tensorflow_model_dataset separately, because these tests are flaky when run as part of the full test-suite
29
- dataset-tests : dataset-doctests
30
- cd kedro-datasets && pytest tests --cov-config pyproject.toml --numprocesses 4 --dist loadfile --ignore tests/tensorflow
31
- cd kedro-datasets && pytest tests/tensorflow/test_tensorflow_model_dataset.py --no-cov
32
-
33
- extra_pytest_args-no-spark =--ignore kedro_datasets/databricks --ignore kedro_datasets/spark
34
- extra_pytest_args =
35
- dataset-doctest% :
36
- if [ " ${* } " != ' s-no-spark' ] && [ " ${* } " != ' s' ]; then \
37
- echo " make: *** No rule to make target \` ${@ } \` . Stop." ; \
38
- exit 2; \
39
- fi ; \
40
- \
41
- # The ignored datasets below require complicated setup with cloud/database clients which is overkill for the doctest examples.
42
- cd kedro-datasets && pytest kedro_datasets --doctest-modules --doctest-continue-on-failure --no-cov \
43
- --ignore kedro_datasets/pandas/gbq_dataset.py \
44
- --ignore kedro_datasets/partitions/partitioned_dataset.py \
45
- --ignore kedro_datasets/redis/redis_dataset.py \
46
- --ignore kedro_datasets/snowflake/snowpark_dataset.py \
47
- --ignore kedro_datasets/spark/spark_hive_dataset.py \
48
- --ignore kedro_datasets/spark/spark_jdbc_dataset.py \
49
- $(extra_pytest_arg${*})
50
-
51
- test-sequential :
52
- cd $(plugin ) && pytest tests --cov-config pyproject.toml
53
-
54
21
e2e-tests :
55
22
cd $(plugin ) && behave
56
23
57
24
secret-scan :
58
25
trufflehog --max_depth 1 --exclude_paths trufflehog-ignore.txt .
59
26
60
- clean :
61
- cd $(plugin ) ; \
62
- rm -rf build dist pip-wheel-metadata .pytest_cache; \
63
- find . -regex " .*/__pycache__" -exec rm -rf {} +; \
64
- find . -regex " .*\.egg-info" -exec rm -rf {} +; \
65
-
66
27
install-test-requirements :
67
- cd $(plugin ) && pip install " .[test]"
28
+ cd $(plugin ) && uv pip install " .[test]"
68
29
69
30
install-pre-commit :
70
31
pre-commit install --install-hooks
@@ -79,12 +40,12 @@ sign-off:
79
40
echo ' --in-place "$$1"' >> .git/hooks/commit-msg
80
41
chmod +x .git/hooks/commit-msg
81
42
43
+ # # kedro-datasets specific
44
+
82
45
# kedro-datasets related only
83
46
test-no-spark : dataset-doctests-no-spark
84
47
cd kedro-datasets && pytest tests --no-cov --ignore tests/spark --ignore tests/databricks --numprocesses 4 --dist loadfile
85
48
86
- test-no-spark-sequential : dataset-doctests-no-spark
87
- cd kedro-datasets && pytest tests --no-cov --ignore tests/spark --ignore tests/databricks
88
49
89
50
# kedro-datasets/snowflake tests skipped from default scope
90
51
test-snowflake-only :
@@ -93,3 +54,26 @@ test-snowflake-only:
93
54
94
55
check-datasets-docs :
95
56
cd kedro-datasets && python -m sphinx -WETan -j auto -D language=en -b linkcheck -d _build/doctrees docs/source _build/linkcheck
57
+
58
+ # Run test_tensorflow_model_dataset separately, because these tests are flaky when run as part of the full test-suite
59
+ dataset-tests : dataset-doctests
60
+ cd kedro-datasets && pytest tests --cov-config pyproject.toml --numprocesses 4 --dist loadfile --ignore tests/tensorflow
61
+ cd kedro-datasets && pytest tests/tensorflow/test_tensorflow_model_dataset.py --no-cov
62
+
63
+ extra_pytest_args-no-spark =--ignore kedro_datasets/databricks --ignore kedro_datasets/spark
64
+ extra_pytest_args =
65
+ dataset-doctest% :
66
+ if [ " ${* } " != ' s-no-spark' ] && [ " ${* } " != ' s' ]; then \
67
+ echo " make: *** No rule to make target \` ${@ } \` . Stop." ; \
68
+ exit 2; \
69
+ fi ; \
70
+ \
71
+ # The ignored datasets below require complicated setup with cloud/database clients which is overkill for the doctest examples.
72
+ cd kedro-datasets && pytest kedro_datasets --doctest-modules --doctest-continue-on-failure --no-cov \
73
+ --ignore kedro_datasets/pandas/gbq_dataset.py \
74
+ --ignore kedro_datasets/partitions/partitioned_dataset.py \
75
+ --ignore kedro_datasets/redis/redis_dataset.py \
76
+ --ignore kedro_datasets/snowflake/snowpark_dataset.py \
77
+ --ignore kedro_datasets/spark/spark_hive_dataset.py \
78
+ --ignore kedro_datasets/spark/spark_jdbc_dataset.py \
79
+ $(extra_pytest_arg${*})
0 commit comments