-
Notifications
You must be signed in to change notification settings - Fork 2
Expand file tree
/
Copy pathpyproject.toml
More file actions
169 lines (154 loc) · 4.69 KB
/
pyproject.toml
File metadata and controls
169 lines (154 loc) · 4.69 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
[build-system]
requires = ["hatchling"]
build-backend = "hatchling.build"
[project]
name = "parallel-web-tools"
version = "0.3.0"
description = "Parallel Tools: CLI and Python SDK for AI-powered web intelligence"
readme = "README.md"
requires-python = ">=3.10"
license = "MIT"
authors = [
{ name = "Parallel", email = "support@parallel.ai" }
]
keywords = ["ai", "data-enrichment", "parallel", "etl", "data-pipeline", "web-search", "llm"]
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.10",
"Programming Language :: Python :: 3.11",
"Programming Language :: Python :: 3.12",
"Programming Language :: Python :: 3.13",
"Programming Language :: Python :: 3.14",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"Topic :: Database",
]
dependencies = [
"parallel-web>=0.6.0",
"python-dotenv>=1.0.0",
# CLI dependencies (minimal - search, extract, enrich with CLI args)
"click>=8.1.0",
"rich>=13.0.0",
"httpx>=0.25.0",
]
[project.urls]
Homepage = "https://github.com/parallel-web/parallel-web-tools"
Documentation = "https://docs.parallel.ai"
Repository = "https://github.com/parallel-web/parallel-web-tools"
Issues = "https://github.com/parallel-web/parallel-web-tools/issues"
[project.scripts]
parallel-cli = "parallel_web_tools.cli:main"
[project.optional-dependencies]
# CLI extras: YAML config files and interactive planner
cli = [
"pyyaml>=6.0.0",
"questionary>=2.0.0",
]
# Polars integration for data processing
polars = [
"polars>=1.37.0",
"pyarrow>=18.0.0",
]
# Pandas integration (for Spark UDFs)
pandas = [
"pandas>=2.3.0",
]
# DuckDB integration (includes polars for data handling)
duckdb = [
"parallel-web-tools[cli,polars]",
"duckdb>=1.0.0",
"nest-asyncio>=1.6.0",
]
# Snowflake integration
snowflake = [
"snowflake-connector-python>=3.0.0",
]
# SQLAlchemy-based BigQuery connector (for YAML configs and ETL)
bigquery = [
"parallel-web-tools[cli]",
"sqlalchemy>=2.0.0",
"sqlalchemy-bigquery>=1.11.0",
]
# BigQuery native UDF deployment (Cloud Functions + Remote Functions)
# Note: Requires gcloud CLI to be installed. See docs/bigquery-setup.md
bigquery-native = [] # No Python deps, just CLI tools
# Apache Spark integration
spark = [
"parallel-web-tools[pandas]",
"pyspark>=3.4.0",
]
# All features (recommended for full CLI experience)
all = [
"parallel-web-tools[cli,polars,duckdb,snowflake,bigquery]",
]
# Development
dev = [
"parallel-web-tools[all,spark]",
"pytest>=9.0.0",
"pytest-cov>=7.0.0",
"pyinstaller>=6.20.0",
"pre-commit>=4.6.0",
"ruff>=0.15.0",
"ty>=0.0.33",
]
[tool.hatch.build.targets.wheel]
packages = ["parallel_web_tools"]
artifacts = [
"*.sql",
"*.txt", # requirements.txt for cloud function
]
[tool.hatch.build.targets.sdist]
include = [
"/parallel_web_tools",
]
[tool.ty.src]
# Exclude files that have external dependencies not in the main project
exclude = [
"**/integrations/bigquery/cloud_function/**", # Separate deployment with flask, functions_framework
"**/integrations/spark/**", # PySpark type stubs are incomplete (pandas_udf overloads)
"**/integrations/duckdb/udf.py", # DuckDB UDF uses type="arrow" not in type stubs
"scripts/runtime_hook_ssl.py", # PyInstaller runtime hook with sys._MEIPASS
"notebooks/**", # Jupyter notebooks with Databricks display()
"examples/**", # Example scripts
]
# SDK returns TaskRunEvent | ErrorEvent union; code narrows via event.type check
# but ty can't follow string-based narrowing on .type attribute
[[tool.ty.overrides]]
include = ["**/core/batch.py"]
[tool.ty.overrides.rules]
unresolved-attribute = "ignore"
# Polars .collect() returns InProcessQuery | DataFrame in newer polars stubs
[[tool.ty.overrides]]
include = ["**/integrations/polars/**"]
[tool.ty.overrides.rules]
invalid-argument-type = "ignore"
[tool.ruff]
line-length = 120
target-version = "py310"
[tool.ruff.lint]
select = [
"E", # pycodestyle errors
"W", # pycodestyle warnings
"F", # pyflakes
"I", # isort
"B", # flake8-bugbear
"UP", # pyupgrade
]
ignore = [
"E501", # line too long (handled by formatter)
"B008", # do not perform function calls in argument defaults
]
[tool.ruff.lint.isort]
known-first-party = ["parallel_web_tools"]
[tool.ruff.lint.per-file-ignores]
"examples/*" = ["E402"] # Allow imports not at top for tutorial scripts
[dependency-groups]
dev = [
"ipykernel>=7.2.0",
"pyinstaller>=6.20.0",
"tach>=0.34.1",
"ty>=0.0.33",
]