-
Notifications
You must be signed in to change notification settings - Fork 39
/
pyproject.toml
107 lines (103 loc) · 3.53 KB
/
pyproject.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
# Copyright (c) 2024, NVIDIA CORPORATION.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
[build-system]
requires = ["setuptools"]
build-backend = "setuptools.build_meta"
[project]
name = "spark-rapids-user-tools"
authors = [
{ name="NVIDIA Corporation" , email="[email protected]" },
]
description = "A simple wrapper process around cloud service providers to run tools for the RAPIDS Accelerator for Apache Spark."
readme = "README.md"
requires-python = ">=3.8,<3.12"
classifiers = [
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.8",
"Programming Language :: Python :: 3.9",
"Programming Language :: Python :: 3.10",
"Programming Language :: Python :: 3.11",
"License :: OSI Approved :: Apache Software License",
"Operating System :: OS Independent",
]
dependencies = [
# numpy-1.24.4 is the latest version to support python3.8. P.S: does not support python-3.12+
"numpy<=1.24.4",
"chevron==0.14.0",
"fastprogress==1.0.3",
"fastcore==1.7.10",
"fire>=0.5.0",
"pandas==1.4.3",
"pyYAML>=6.0,<=7.0",
# This is used to resolve env-variable in yaml files. It requires netween 5.0 and 6.0
"pyaml-env==1.2.1",
"tabulate==0.8.10",
"importlib-resources==5.10.2",
"requests==2.31.0",
"packaging>=23.0",
"certifi==2024.7.4",
"idna==3.4",
"urllib3==1.26.19",
"beautifulsoup4==4.11.2",
"pygments==2.15.0",
# used to apply validator on objects and models. "2.9.2" contains from_json method.
"pydantic==2.9.2",
# used to help pylint understand pydantic
"pylint-pydantic==0.3.0",
# used for common API to access remote filesystems like local/s3/gcs/hdfs
# pin to 16.1.0 which works for both numpy-1.0 and numpy-2.0
"pyarrow==16.1.0",
# used for ADLS filesystem implementation
# Issue-568: use 12.17.0 as the new 12.18.0 causes an error in runtime
"azure-storage-blob==12.17.0",
"adlfs==2023.4.0",
# used for spinner animation
"progress==1.6",
# used for model estimations
"xgboost==2.0.3",
# used for model interpretability
"shap==0.44.1",
# used for retrieving available memory on the host
"psutil==5.9.8"
]
dynamic=["entry-points", "version"]
[project.scripts]
spark_rapids_user_tools = "spark_rapids_pytools.wrapper:main"
spark_rapids = "spark_rapids_tools.cmdli.tools_cli:main"
spark_rapids_dev = "spark_rapids_tools.cmdli.dev_cli:main"
[tool.setuptools]
package-dir = {"" = "src"}
[tool.setuptools.packages.find]
where = ["src"]
[tool.setuptools.dynamic]
version = {attr = "spark_rapids_pytools.__version__"}
[tool.setuptools.package-data]
"*"= ["*.json", "*.yaml", "*.ms", "*.sh", "*.tgz", "*.properties"]
[tool.poetry]
repository = "https://github.com/NVIDIA/spark-rapids-tools/tree/main"
[project.optional-dependencies]
test = [
"tox", 'pytest', 'cli_test_helpers', 'behave',
# use flak-8 plugin for pydantic
'flake8-pydantic',
# use pylint specific version
'pylint==3.2.7',
]
qualx = [
"holoviews",
"matplotlib",
"optuna",
"optuna-integration",
"seaborn"
]