Compare commits
35 Commits
fix/pypi-p
...
3c2ab70ef0
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
3c2ab70ef0 | ||
|
|
109a593f83 | ||
|
|
1d727b3f8c | ||
|
|
47aa0afe91 | ||
|
|
f992fc540e | ||
|
|
044a6c1d27 | ||
|
|
62c77dc16d | ||
|
|
7c05360eed | ||
|
|
76878279e9 | ||
|
|
e1b01abf9b | ||
|
|
d07936b666 | ||
|
|
47b3eb439d | ||
|
|
c5f75e4fd6 | ||
|
|
ff31379649 | ||
|
|
424b1e5770 | ||
|
|
7b5b0c78d8 | ||
|
|
924826f07a | ||
|
|
fe6c6c52d2 | ||
|
|
701e11ce83 | ||
|
|
ff9e02606e | ||
|
|
f3afdd3bbf | ||
|
|
4b73196664 | ||
|
|
7ef66745f1 | ||
|
|
2dc7fe5a7b | ||
|
|
534e4b964f | ||
|
|
757e43fc34 | ||
|
|
d78092de55 | ||
|
|
0fa991f536 | ||
|
|
00fb2729e4 | ||
|
|
8ae4d7a685 | ||
|
|
4b887d1aad | ||
|
|
4dc54ace8a | ||
|
|
64bfd3902f | ||
|
|
bdfed77cb1 | ||
|
|
140f6c926a |
@@ -10,11 +10,20 @@ Handles:
|
|||||||
- Conflict detection
|
- Conflict detection
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
import re
|
||||||
import yaml
|
import yaml
|
||||||
from typing import List, Dict, Any, Optional, Set, Tuple
|
from typing import List, Dict, Any, Optional, Set, Tuple
|
||||||
from sqlalchemy.orm import Session
|
from sqlalchemy.orm import Session
|
||||||
from sqlalchemy import and_
|
from sqlalchemy import and_
|
||||||
|
|
||||||
|
# Import packaging for PEP 440 version matching
|
||||||
|
try:
|
||||||
|
from packaging.specifiers import SpecifierSet, InvalidSpecifier
|
||||||
|
from packaging.version import Version, InvalidVersion
|
||||||
|
HAS_PACKAGING = True
|
||||||
|
except ImportError:
|
||||||
|
HAS_PACKAGING = False
|
||||||
|
|
||||||
from .models import (
|
from .models import (
|
||||||
Project,
|
Project,
|
||||||
Package,
|
Package,
|
||||||
@@ -304,6 +313,87 @@ def get_reverse_dependencies(
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def _is_version_constraint(version_str: str) -> bool:
|
||||||
|
"""Check if a version string contains constraint operators."""
|
||||||
|
if not version_str:
|
||||||
|
return False
|
||||||
|
# Check for common constraint operators
|
||||||
|
return any(op in version_str for op in ['>=', '<=', '!=', '~=', '>', '<', '==', '*'])
|
||||||
|
|
||||||
|
|
||||||
|
def _resolve_version_constraint(
|
||||||
|
db: Session,
|
||||||
|
package: Package,
|
||||||
|
constraint: str,
|
||||||
|
) -> Optional[Tuple[str, str, int]]:
|
||||||
|
"""
|
||||||
|
Resolve a version constraint (e.g., '>=1.9') to a specific version.
|
||||||
|
|
||||||
|
Uses PEP 440 version matching to find the best matching version.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
db: Database session
|
||||||
|
package: Package to search versions in
|
||||||
|
constraint: Version constraint string (e.g., '>=1.9', '<2.0,>=1.5')
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tuple of (artifact_id, resolved_version, size) or None if not found
|
||||||
|
"""
|
||||||
|
if not HAS_PACKAGING:
|
||||||
|
# Fallback: if packaging not available, can't do constraint matching
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Handle wildcard - return latest version
|
||||||
|
if constraint == '*':
|
||||||
|
# Get the latest version by created_at
|
||||||
|
latest = db.query(PackageVersion).filter(
|
||||||
|
PackageVersion.package_id == package.id,
|
||||||
|
).order_by(PackageVersion.created_at.desc()).first()
|
||||||
|
if latest:
|
||||||
|
artifact = db.query(Artifact).filter(Artifact.id == latest.artifact_id).first()
|
||||||
|
if artifact:
|
||||||
|
return (artifact.id, latest.version, artifact.size)
|
||||||
|
return None
|
||||||
|
|
||||||
|
try:
|
||||||
|
specifier = SpecifierSet(constraint)
|
||||||
|
except InvalidSpecifier:
|
||||||
|
# Invalid constraint, try as exact version
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Get all versions for this package
|
||||||
|
all_versions = db.query(PackageVersion).filter(
|
||||||
|
PackageVersion.package_id == package.id,
|
||||||
|
).all()
|
||||||
|
|
||||||
|
if not all_versions:
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Find matching versions
|
||||||
|
matching = []
|
||||||
|
for pv in all_versions:
|
||||||
|
try:
|
||||||
|
v = Version(pv.version)
|
||||||
|
if v in specifier:
|
||||||
|
matching.append((pv, v))
|
||||||
|
except InvalidVersion:
|
||||||
|
# Skip invalid versions
|
||||||
|
continue
|
||||||
|
|
||||||
|
if not matching:
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Sort by version (descending) and return the latest matching
|
||||||
|
matching.sort(key=lambda x: x[1], reverse=True)
|
||||||
|
best_match = matching[0][0]
|
||||||
|
|
||||||
|
artifact = db.query(Artifact).filter(Artifact.id == best_match.artifact_id).first()
|
||||||
|
if artifact:
|
||||||
|
return (artifact.id, best_match.version, artifact.size)
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
def _resolve_dependency_to_artifact(
|
def _resolve_dependency_to_artifact(
|
||||||
db: Session,
|
db: Session,
|
||||||
project_name: str,
|
project_name: str,
|
||||||
@@ -314,11 +404,17 @@ def _resolve_dependency_to_artifact(
|
|||||||
"""
|
"""
|
||||||
Resolve a dependency constraint to an artifact ID.
|
Resolve a dependency constraint to an artifact ID.
|
||||||
|
|
||||||
|
Supports:
|
||||||
|
- Exact version matching (e.g., '1.2.3')
|
||||||
|
- Version constraints (e.g., '>=1.9', '<2.0,>=1.5')
|
||||||
|
- Tag matching
|
||||||
|
- Wildcard ('*' for any version)
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
db: Database session
|
db: Database session
|
||||||
project_name: Project name
|
project_name: Project name
|
||||||
package_name: Package name
|
package_name: Package name
|
||||||
version: Version constraint (exact)
|
version: Version or version constraint
|
||||||
tag: Tag constraint
|
tag: Tag constraint
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
@@ -337,17 +433,23 @@ def _resolve_dependency_to_artifact(
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
if version:
|
if version:
|
||||||
# Look up by version
|
# Check if this is a version constraint (>=, <, etc.) or exact version
|
||||||
pkg_version = db.query(PackageVersion).filter(
|
if _is_version_constraint(version):
|
||||||
PackageVersion.package_id == package.id,
|
result = _resolve_version_constraint(db, package, version)
|
||||||
PackageVersion.version == version,
|
if result:
|
||||||
).first()
|
return result
|
||||||
if pkg_version:
|
else:
|
||||||
artifact = db.query(Artifact).filter(
|
# Look up by exact version
|
||||||
Artifact.id == pkg_version.artifact_id
|
pkg_version = db.query(PackageVersion).filter(
|
||||||
|
PackageVersion.package_id == package.id,
|
||||||
|
PackageVersion.version == version,
|
||||||
).first()
|
).first()
|
||||||
if artifact:
|
if pkg_version:
|
||||||
return (artifact.id, version, artifact.size)
|
artifact = db.query(Artifact).filter(
|
||||||
|
Artifact.id == pkg_version.artifact_id
|
||||||
|
).first()
|
||||||
|
if artifact:
|
||||||
|
return (artifact.id, version, artifact.size)
|
||||||
|
|
||||||
# Also check if there's a tag with this exact name
|
# Also check if there's a tag with this exact name
|
||||||
tag_record = db.query(Tag).filter(
|
tag_record = db.query(Tag).filter(
|
||||||
|
|||||||
@@ -8,35 +8,196 @@ Artifacts are cached on first access through configured upstream sources.
|
|||||||
import hashlib
|
import hashlib
|
||||||
import logging
|
import logging
|
||||||
import re
|
import re
|
||||||
from typing import Optional
|
import tarfile
|
||||||
|
import threading
|
||||||
|
import zipfile
|
||||||
|
from io import BytesIO
|
||||||
|
from typing import Optional, List, Tuple
|
||||||
from urllib.parse import urljoin, urlparse, quote, unquote
|
from urllib.parse import urljoin, urlparse, quote, unquote
|
||||||
|
|
||||||
import httpx
|
import httpx
|
||||||
from fastapi import APIRouter, Depends, HTTPException, Request, Response
|
from fastapi import APIRouter, BackgroundTasks, Depends, HTTPException, Request, Response
|
||||||
from fastapi.responses import StreamingResponse, HTMLResponse
|
from fastapi.responses import StreamingResponse, HTMLResponse
|
||||||
from sqlalchemy.orm import Session
|
from sqlalchemy.orm import Session
|
||||||
|
|
||||||
from .database import get_db
|
from .database import get_db
|
||||||
from .models import UpstreamSource, CachedUrl, Artifact, Project, Package, Tag
|
from .models import UpstreamSource, CachedUrl, Artifact, Project, Package, Tag, PackageVersion, ArtifactDependency
|
||||||
from .storage import S3Storage, get_storage
|
from .storage import S3Storage, get_storage
|
||||||
from .upstream import (
|
|
||||||
UpstreamClient,
|
|
||||||
UpstreamClientConfig,
|
|
||||||
UpstreamHTTPError,
|
|
||||||
UpstreamConnectionError,
|
|
||||||
UpstreamTimeoutError,
|
|
||||||
)
|
|
||||||
from .config import get_env_upstream_sources
|
from .config import get_env_upstream_sources
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
router = APIRouter(prefix="/pypi", tags=["pypi-proxy"])
|
router = APIRouter(prefix="/pypi", tags=["pypi-proxy"])
|
||||||
|
|
||||||
|
|
||||||
|
def _parse_requires_dist(requires_dist: str) -> Tuple[str, Optional[str]]:
|
||||||
|
"""Parse a Requires-Dist line into (package_name, version_constraint).
|
||||||
|
|
||||||
|
Examples:
|
||||||
|
"requests (>=2.25.0)" -> ("requests", ">=2.25.0")
|
||||||
|
"typing-extensions; python_version < '3.8'" -> ("typing-extensions", None)
|
||||||
|
"numpy>=1.21.0" -> ("numpy", ">=1.21.0")
|
||||||
|
"certifi" -> ("certifi", None)
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tuple of (normalized_package_name, version_constraint or None)
|
||||||
|
"""
|
||||||
|
# Remove any environment markers (after semicolon)
|
||||||
|
if ';' in requires_dist:
|
||||||
|
requires_dist = requires_dist.split(';')[0].strip()
|
||||||
|
|
||||||
|
# Match patterns like "package (>=1.0)" or "package>=1.0" or "package"
|
||||||
|
# Pattern breakdown: package name, optional whitespace, optional version in parens or directly
|
||||||
|
match = re.match(
|
||||||
|
r'^([a-zA-Z0-9][-a-zA-Z0-9._]*)\s*(?:\(([^)]+)\)|([<>=!~][^\s;]+))?',
|
||||||
|
requires_dist.strip()
|
||||||
|
)
|
||||||
|
|
||||||
|
if not match:
|
||||||
|
return None, None
|
||||||
|
|
||||||
|
package_name = match.group(1)
|
||||||
|
# Version can be in parentheses (group 2) or directly after name (group 3)
|
||||||
|
version_constraint = match.group(2) or match.group(3)
|
||||||
|
|
||||||
|
# Normalize package name (PEP 503)
|
||||||
|
normalized_name = re.sub(r'[-_.]+', '-', package_name).lower()
|
||||||
|
|
||||||
|
# Clean up version constraint
|
||||||
|
if version_constraint:
|
||||||
|
version_constraint = version_constraint.strip()
|
||||||
|
|
||||||
|
return normalized_name, version_constraint
|
||||||
|
|
||||||
|
|
||||||
|
def _extract_requires_from_metadata(metadata_content: str) -> List[Tuple[str, Optional[str]]]:
|
||||||
|
"""Extract all Requires-Dist entries from METADATA/PKG-INFO content.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
metadata_content: The content of a METADATA or PKG-INFO file
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
List of (package_name, version_constraint) tuples
|
||||||
|
"""
|
||||||
|
dependencies = []
|
||||||
|
|
||||||
|
for line in metadata_content.split('\n'):
|
||||||
|
if line.startswith('Requires-Dist:'):
|
||||||
|
# Extract the value after "Requires-Dist:"
|
||||||
|
value = line[len('Requires-Dist:'):].strip()
|
||||||
|
pkg_name, version = _parse_requires_dist(value)
|
||||||
|
if pkg_name:
|
||||||
|
dependencies.append((pkg_name, version))
|
||||||
|
|
||||||
|
return dependencies
|
||||||
|
|
||||||
|
|
||||||
|
def _extract_metadata_from_wheel(content: bytes) -> Optional[str]:
|
||||||
|
"""Extract METADATA file content from a wheel (zip) file.
|
||||||
|
|
||||||
|
Wheel files have structure: {package}-{version}.dist-info/METADATA
|
||||||
|
|
||||||
|
Args:
|
||||||
|
content: The wheel file content as bytes
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
METADATA file content as string, or None if not found
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
with zipfile.ZipFile(BytesIO(content)) as zf:
|
||||||
|
# Find the .dist-info directory
|
||||||
|
for name in zf.namelist():
|
||||||
|
if name.endswith('.dist-info/METADATA'):
|
||||||
|
return zf.read(name).decode('utf-8', errors='replace')
|
||||||
|
except Exception as e:
|
||||||
|
logger.warning(f"Failed to extract metadata from wheel: {e}")
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def _extract_metadata_from_sdist(content: bytes, filename: str) -> Optional[str]:
|
||||||
|
"""Extract PKG-INFO file content from a source distribution (.tar.gz).
|
||||||
|
|
||||||
|
Source distributions have structure: {package}-{version}/PKG-INFO
|
||||||
|
|
||||||
|
Args:
|
||||||
|
content: The tarball content as bytes
|
||||||
|
filename: The original filename (used to determine package name)
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
PKG-INFO file content as string, or None if not found
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
with tarfile.open(fileobj=BytesIO(content), mode='r:gz') as tf:
|
||||||
|
# Find PKG-INFO in the root directory of the archive
|
||||||
|
for member in tf.getmembers():
|
||||||
|
if member.name.endswith('/PKG-INFO') and member.name.count('/') == 1:
|
||||||
|
f = tf.extractfile(member)
|
||||||
|
if f:
|
||||||
|
return f.read().decode('utf-8', errors='replace')
|
||||||
|
except Exception as e:
|
||||||
|
logger.warning(f"Failed to extract metadata from sdist {filename}: {e}")
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def _extract_dependencies(content: bytes, filename: str) -> List[Tuple[str, Optional[str]]]:
|
||||||
|
"""Extract dependencies from a PyPI package file.
|
||||||
|
|
||||||
|
Supports wheel (.whl) and source distribution (.tar.gz) formats.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
content: The package file content as bytes
|
||||||
|
filename: The original filename
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
List of (package_name, version_constraint) tuples
|
||||||
|
"""
|
||||||
|
metadata = None
|
||||||
|
|
||||||
|
if filename.endswith('.whl'):
|
||||||
|
metadata = _extract_metadata_from_wheel(content)
|
||||||
|
elif filename.endswith('.tar.gz'):
|
||||||
|
metadata = _extract_metadata_from_sdist(content, filename)
|
||||||
|
|
||||||
|
if metadata:
|
||||||
|
return _extract_requires_from_metadata(metadata)
|
||||||
|
|
||||||
|
return []
|
||||||
|
|
||||||
# Timeout configuration for proxy requests
|
# Timeout configuration for proxy requests
|
||||||
PROXY_CONNECT_TIMEOUT = 30.0
|
PROXY_CONNECT_TIMEOUT = 30.0
|
||||||
PROXY_READ_TIMEOUT = 60.0
|
PROXY_READ_TIMEOUT = 60.0
|
||||||
|
|
||||||
|
|
||||||
|
def _extract_pypi_version(filename: str) -> Optional[str]:
|
||||||
|
"""Extract version from PyPI filename.
|
||||||
|
|
||||||
|
Handles formats like:
|
||||||
|
- cowsay-6.1-py3-none-any.whl
|
||||||
|
- cowsay-1.0.tar.gz
|
||||||
|
- some_package-1.2.3.post1-cp39-cp39-linux_x86_64.whl
|
||||||
|
"""
|
||||||
|
# Remove extension
|
||||||
|
if filename.endswith('.whl'):
|
||||||
|
# Wheel: name-version-pytag-abitag-platform.whl
|
||||||
|
parts = filename[:-4].split('-')
|
||||||
|
if len(parts) >= 2:
|
||||||
|
return parts[1]
|
||||||
|
elif filename.endswith('.tar.gz'):
|
||||||
|
# Source: name-version.tar.gz
|
||||||
|
base = filename[:-7]
|
||||||
|
# Find the last hyphen that precedes a version-like string
|
||||||
|
match = re.match(r'^(.+)-(\d+.*)$', base)
|
||||||
|
if match:
|
||||||
|
return match.group(2)
|
||||||
|
elif filename.endswith('.zip'):
|
||||||
|
# Egg/zip: name-version.zip
|
||||||
|
base = filename[:-4]
|
||||||
|
match = re.match(r'^(.+)-(\d+.*)$', base)
|
||||||
|
if match:
|
||||||
|
return match.group(2)
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
def _get_pypi_upstream_sources(db: Session) -> list[UpstreamSource]:
|
def _get_pypi_upstream_sources(db: Session) -> list[UpstreamSource]:
|
||||||
"""Get all enabled upstream sources configured for PyPI."""
|
"""Get all enabled upstream sources configured for PyPI."""
|
||||||
# Get database sources
|
# Get database sources
|
||||||
@@ -88,7 +249,27 @@ def _get_basic_auth(source) -> Optional[tuple[str, str]]:
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
def _rewrite_package_links(html: str, base_url: str, package_name: str) -> str:
|
def _get_base_url(request: Request) -> str:
|
||||||
|
"""
|
||||||
|
Get the external base URL, respecting X-Forwarded-Proto header.
|
||||||
|
|
||||||
|
When behind a reverse proxy that terminates SSL, the request.base_url
|
||||||
|
will show http:// even though the external URL is https://. This function
|
||||||
|
checks the X-Forwarded-Proto header to determine the correct scheme.
|
||||||
|
"""
|
||||||
|
base_url = str(request.base_url).rstrip('/')
|
||||||
|
|
||||||
|
# Check for X-Forwarded-Proto header (set by reverse proxies)
|
||||||
|
forwarded_proto = request.headers.get('x-forwarded-proto')
|
||||||
|
if forwarded_proto:
|
||||||
|
# Replace the scheme with the forwarded protocol
|
||||||
|
parsed = urlparse(base_url)
|
||||||
|
base_url = f"{forwarded_proto}://{parsed.netloc}{parsed.path}"
|
||||||
|
|
||||||
|
return base_url
|
||||||
|
|
||||||
|
|
||||||
|
def _rewrite_package_links(html: str, base_url: str, package_name: str, upstream_base_url: str) -> str:
|
||||||
"""
|
"""
|
||||||
Rewrite download links in a PyPI simple page to go through our proxy.
|
Rewrite download links in a PyPI simple page to go through our proxy.
|
||||||
|
|
||||||
@@ -96,6 +277,7 @@ def _rewrite_package_links(html: str, base_url: str, package_name: str) -> str:
|
|||||||
html: The HTML content from upstream
|
html: The HTML content from upstream
|
||||||
base_url: Our server's base URL
|
base_url: Our server's base URL
|
||||||
package_name: The package name for the URL path
|
package_name: The package name for the URL path
|
||||||
|
upstream_base_url: The upstream URL used to fetch this page (for resolving relative URLs)
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
HTML with rewritten download links
|
HTML with rewritten download links
|
||||||
@@ -103,19 +285,31 @@ def _rewrite_package_links(html: str, base_url: str, package_name: str) -> str:
|
|||||||
# Pattern to match href attributes in anchor tags
|
# Pattern to match href attributes in anchor tags
|
||||||
# PyPI simple pages have links like:
|
# PyPI simple pages have links like:
|
||||||
# <a href="https://files.pythonhosted.org/packages/.../file.tar.gz#sha256=...">file.tar.gz</a>
|
# <a href="https://files.pythonhosted.org/packages/.../file.tar.gz#sha256=...">file.tar.gz</a>
|
||||||
|
# Or relative URLs from Artifactory like:
|
||||||
|
# <a href="../../packages/packages/62/35/.../requests-0.10.0.tar.gz#sha256=...">
|
||||||
|
|
||||||
def replace_href(match):
|
def replace_href(match):
|
||||||
original_url = match.group(1)
|
original_url = match.group(1)
|
||||||
|
|
||||||
|
# Resolve relative URLs to absolute using the upstream base URL
|
||||||
|
if not original_url.startswith(('http://', 'https://')):
|
||||||
|
# Split off fragment before resolving
|
||||||
|
url_without_fragment = original_url.split('#')[0]
|
||||||
|
fragment_part = original_url[len(url_without_fragment):]
|
||||||
|
absolute_url = urljoin(upstream_base_url, url_without_fragment) + fragment_part
|
||||||
|
else:
|
||||||
|
absolute_url = original_url
|
||||||
|
|
||||||
# Extract the filename from the URL
|
# Extract the filename from the URL
|
||||||
parsed = urlparse(original_url)
|
parsed = urlparse(absolute_url)
|
||||||
path_parts = parsed.path.split('/')
|
path_parts = parsed.path.split('/')
|
||||||
filename = path_parts[-1] if path_parts else ''
|
filename = path_parts[-1] if path_parts else ''
|
||||||
|
|
||||||
# Keep the hash fragment if present
|
# Keep the hash fragment if present
|
||||||
fragment = f"#{parsed.fragment}" if parsed.fragment else ""
|
fragment = f"#{parsed.fragment}" if parsed.fragment else ""
|
||||||
|
|
||||||
# Encode the original URL for safe transmission
|
# Encode the absolute URL (without fragment) for safe transmission
|
||||||
encoded_url = quote(original_url.split('#')[0], safe='')
|
encoded_url = quote(absolute_url.split('#')[0], safe='')
|
||||||
|
|
||||||
# Build new URL pointing to our proxy
|
# Build new URL pointing to our proxy
|
||||||
new_url = f"{base_url}/pypi/simple/{package_name}/{filename}?upstream={encoded_url}{fragment}"
|
new_url = f"{base_url}/pypi/simple/{package_name}/{filename}?upstream={encoded_url}{fragment}"
|
||||||
@@ -154,12 +348,10 @@ async def pypi_simple_index(
|
|||||||
headers.update(_build_auth_headers(source))
|
headers.update(_build_auth_headers(source))
|
||||||
auth = _get_basic_auth(source)
|
auth = _get_basic_auth(source)
|
||||||
|
|
||||||
simple_url = source.url.rstrip('/') + '/simple/'
|
# Use URL as-is - users should provide full path including /simple
|
||||||
|
simple_url = source.url.rstrip('/') + '/'
|
||||||
|
|
||||||
timeout = httpx.Timeout(
|
timeout = httpx.Timeout(PROXY_READ_TIMEOUT, connect=PROXY_CONNECT_TIMEOUT)
|
||||||
connect=PROXY_CONNECT_TIMEOUT,
|
|
||||||
read=PROXY_READ_TIMEOUT,
|
|
||||||
)
|
|
||||||
|
|
||||||
with httpx.Client(timeout=timeout, follow_redirects=False) as client:
|
with httpx.Client(timeout=timeout, follow_redirects=False) as client:
|
||||||
response = client.get(
|
response = client.get(
|
||||||
@@ -186,7 +378,7 @@ async def pypi_simple_index(
|
|||||||
content = response.text
|
content = response.text
|
||||||
|
|
||||||
# Rewrite package links to go through our proxy
|
# Rewrite package links to go through our proxy
|
||||||
base_url = str(request.base_url).rstrip('/')
|
base_url = _get_base_url(request)
|
||||||
content = re.sub(
|
content = re.sub(
|
||||||
r'href="([^"]+)/"',
|
r'href="([^"]+)/"',
|
||||||
lambda m: f'href="{base_url}/pypi/simple/{m.group(1)}/"',
|
lambda m: f'href="{base_url}/pypi/simple/{m.group(1)}/"',
|
||||||
@@ -232,7 +424,7 @@ async def pypi_package_versions(
|
|||||||
detail="No PyPI upstream sources configured"
|
detail="No PyPI upstream sources configured"
|
||||||
)
|
)
|
||||||
|
|
||||||
base_url = str(request.base_url).rstrip('/')
|
base_url = _get_base_url(request)
|
||||||
|
|
||||||
# Normalize package name (PEP 503)
|
# Normalize package name (PEP 503)
|
||||||
normalized_name = re.sub(r'[-_.]+', '-', package_name).lower()
|
normalized_name = re.sub(r'[-_.]+', '-', package_name).lower()
|
||||||
@@ -245,12 +437,11 @@ async def pypi_package_versions(
|
|||||||
headers.update(_build_auth_headers(source))
|
headers.update(_build_auth_headers(source))
|
||||||
auth = _get_basic_auth(source)
|
auth = _get_basic_auth(source)
|
||||||
|
|
||||||
package_url = source.url.rstrip('/') + f'/simple/{normalized_name}/'
|
# Use URL as-is - users should provide full path including /simple
|
||||||
|
package_url = source.url.rstrip('/') + f'/{normalized_name}/'
|
||||||
|
final_url = package_url # Track final URL after redirects
|
||||||
|
|
||||||
timeout = httpx.Timeout(
|
timeout = httpx.Timeout(PROXY_READ_TIMEOUT, connect=PROXY_CONNECT_TIMEOUT)
|
||||||
connect=PROXY_CONNECT_TIMEOUT,
|
|
||||||
read=PROXY_READ_TIMEOUT,
|
|
||||||
)
|
|
||||||
|
|
||||||
with httpx.Client(timeout=timeout, follow_redirects=False) as client:
|
with httpx.Client(timeout=timeout, follow_redirects=False) as client:
|
||||||
response = client.get(
|
response = client.get(
|
||||||
@@ -268,7 +459,9 @@ async def pypi_package_versions(
|
|||||||
|
|
||||||
# Make redirect URL absolute if needed
|
# Make redirect URL absolute if needed
|
||||||
if not redirect_url.startswith('http'):
|
if not redirect_url.startswith('http'):
|
||||||
redirect_url = urljoin(package_url, redirect_url)
|
redirect_url = urljoin(final_url, redirect_url)
|
||||||
|
|
||||||
|
final_url = redirect_url # Update final URL
|
||||||
|
|
||||||
response = client.get(
|
response = client.get(
|
||||||
redirect_url,
|
redirect_url,
|
||||||
@@ -282,7 +475,8 @@ async def pypi_package_versions(
|
|||||||
content = response.text
|
content = response.text
|
||||||
|
|
||||||
# Rewrite download links to go through our proxy
|
# Rewrite download links to go through our proxy
|
||||||
content = _rewrite_package_links(content, base_url, normalized_name)
|
# Pass final_url so relative URLs can be resolved correctly
|
||||||
|
content = _rewrite_package_links(content, base_url, normalized_name, final_url)
|
||||||
|
|
||||||
return HTMLResponse(content=content)
|
return HTMLResponse(content=content)
|
||||||
|
|
||||||
@@ -309,6 +503,93 @@ async def pypi_package_versions(
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def _cache_dependency_background(
|
||||||
|
base_url: str,
|
||||||
|
dep_name: str,
|
||||||
|
dep_version: Optional[str],
|
||||||
|
depth: int = 0,
|
||||||
|
max_depth: int = 10,
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Background task to proactively cache a dependency.
|
||||||
|
|
||||||
|
Fetches the dependency from upstream via our own proxy, which will
|
||||||
|
recursively cache its dependencies as well.
|
||||||
|
"""
|
||||||
|
if depth >= max_depth:
|
||||||
|
logger.warning(f"PyPI proxy: max depth {max_depth} reached caching {dep_name}")
|
||||||
|
return
|
||||||
|
|
||||||
|
try:
|
||||||
|
# Normalize package name for URL (PEP 503)
|
||||||
|
normalized_name = re.sub(r'[-_.]+', '-', dep_name).lower()
|
||||||
|
|
||||||
|
# First, get the simple index page to find available versions
|
||||||
|
# Use HTTPS explicitly to avoid redirect issues that can drop trailing slashes
|
||||||
|
if base_url.startswith('http://'):
|
||||||
|
base_url = 'https://' + base_url[7:]
|
||||||
|
simple_url = f"{base_url}/pypi/simple/{normalized_name}/"
|
||||||
|
logger.info(f"PyPI proxy: proactively caching {dep_name} (depth={depth})")
|
||||||
|
|
||||||
|
with httpx.Client(timeout=30.0, follow_redirects=True) as client:
|
||||||
|
response = client.get(simple_url)
|
||||||
|
if response.status_code != 200:
|
||||||
|
logger.warning(f"PyPI proxy: failed to get index for {dep_name}: {response.status_code}")
|
||||||
|
return
|
||||||
|
|
||||||
|
# Parse the HTML to find wheel files
|
||||||
|
html = response.text
|
||||||
|
|
||||||
|
# Create pattern that matches both normalized (hyphens) and original (underscores) forms
|
||||||
|
# PEP 503 normalizes to hyphens, but wheel filenames may use underscores
|
||||||
|
name_pattern = re.sub(r'[-_]+', '[-_]+', normalized_name)
|
||||||
|
|
||||||
|
# Look for wheel files (.whl) - prefer them over sdist
|
||||||
|
wheel_pattern = rf'href="([^"]*{name_pattern}[^"]*\.whl[^"]*)"'
|
||||||
|
matches = re.findall(wheel_pattern, html, re.IGNORECASE)
|
||||||
|
|
||||||
|
if not matches:
|
||||||
|
# Try sdist
|
||||||
|
sdist_pattern = rf'href="([^"]*{name_pattern}[^"]*\.tar\.gz[^"]*)"'
|
||||||
|
matches = re.findall(sdist_pattern, html, re.IGNORECASE)
|
||||||
|
|
||||||
|
if not matches:
|
||||||
|
# Debug: log first 500 chars of HTML and the pattern we're looking for
|
||||||
|
logger.warning(f"PyPI proxy: no downloadable files found for {dep_name}. Pattern: {wheel_pattern}, HTML preview: {html[:500]}")
|
||||||
|
return
|
||||||
|
|
||||||
|
# Get the last match (usually the latest version)
|
||||||
|
# The URL might be relative or absolute
|
||||||
|
download_url = matches[-1]
|
||||||
|
if download_url.startswith('/'):
|
||||||
|
download_url = f"{base_url}{download_url}"
|
||||||
|
elif not download_url.startswith('http'):
|
||||||
|
download_url = f"{base_url}/pypi/simple/{normalized_name}/{download_url}"
|
||||||
|
|
||||||
|
# Download the file through our proxy (this will cache it)
|
||||||
|
logger.info(f"PyPI proxy: downloading dependency {dep_name} from {download_url}")
|
||||||
|
response = client.get(download_url)
|
||||||
|
if response.status_code == 200:
|
||||||
|
logger.info(f"PyPI proxy: successfully cached {dep_name}")
|
||||||
|
else:
|
||||||
|
logger.warning(f"PyPI proxy: failed to cache {dep_name}: {response.status_code}")
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
logger.warning(f"PyPI proxy: error caching dependency {dep_name}: {e}")
|
||||||
|
|
||||||
|
|
||||||
|
def _start_background_dependency_caching(base_url: str, dependencies: List[Tuple[str, Optional[str]]]):
|
||||||
|
"""Start background threads to cache dependencies."""
|
||||||
|
for dep_name, dep_version in dependencies:
|
||||||
|
# Use a thread to avoid blocking
|
||||||
|
thread = threading.Thread(
|
||||||
|
target=_cache_dependency_background,
|
||||||
|
args=(base_url, dep_name, dep_version),
|
||||||
|
daemon=True,
|
||||||
|
)
|
||||||
|
thread.start()
|
||||||
|
|
||||||
|
|
||||||
@router.get("/simple/{package_name}/{filename}")
|
@router.get("/simple/{package_name}/{filename}")
|
||||||
async def pypi_download_file(
|
async def pypi_download_file(
|
||||||
request: Request,
|
request: Request,
|
||||||
@@ -347,14 +628,22 @@ async def pypi_download_file(
|
|||||||
|
|
||||||
# Stream from S3
|
# Stream from S3
|
||||||
try:
|
try:
|
||||||
content_stream = storage.get_artifact_stream(artifact.id)
|
stream, content_length, _ = storage.get_stream(artifact.s3_key)
|
||||||
|
|
||||||
|
def stream_content():
|
||||||
|
"""Generator that yields chunks from the S3 stream."""
|
||||||
|
try:
|
||||||
|
for chunk in stream.iter_chunks():
|
||||||
|
yield chunk
|
||||||
|
finally:
|
||||||
|
stream.close()
|
||||||
|
|
||||||
return StreamingResponse(
|
return StreamingResponse(
|
||||||
content_stream,
|
stream_content(),
|
||||||
media_type=artifact.content_type or "application/octet-stream",
|
media_type=artifact.content_type or "application/octet-stream",
|
||||||
headers={
|
headers={
|
||||||
"Content-Disposition": f'attachment; filename="{filename}"',
|
"Content-Disposition": f'attachment; filename="{filename}"',
|
||||||
"Content-Length": str(artifact.size),
|
"Content-Length": str(content_length),
|
||||||
"X-Checksum-SHA256": artifact.id,
|
"X-Checksum-SHA256": artifact.id,
|
||||||
"X-Cache": "HIT",
|
"X-Cache": "HIT",
|
||||||
}
|
}
|
||||||
@@ -366,18 +655,10 @@ async def pypi_download_file(
|
|||||||
# Not cached - fetch from upstream
|
# Not cached - fetch from upstream
|
||||||
sources = _get_pypi_upstream_sources(db)
|
sources = _get_pypi_upstream_sources(db)
|
||||||
|
|
||||||
# Find a source that matches the upstream URL
|
# Use the first available source for authentication headers
|
||||||
matched_source = None
|
# Note: The upstream URL may point to files.pythonhosted.org or other CDNs,
|
||||||
for source in sources:
|
# not the configured source URL directly, so we can't strictly validate the host
|
||||||
source_url = getattr(source, 'url', '')
|
matched_source = sources[0] if sources else None
|
||||||
# Check if the upstream URL could come from this source
|
|
||||||
# (This is a loose check - the URL might be from files.pythonhosted.org)
|
|
||||||
if urlparse(upstream_url).netloc in source_url or True: # Allow any source for now
|
|
||||||
matched_source = source
|
|
||||||
break
|
|
||||||
|
|
||||||
if not matched_source and sources:
|
|
||||||
matched_source = sources[0] # Use first source for auth if available
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
headers = {"User-Agent": "Orchard-PyPI-Proxy/1.0"}
|
headers = {"User-Agent": "Orchard-PyPI-Proxy/1.0"}
|
||||||
@@ -385,10 +666,7 @@ async def pypi_download_file(
|
|||||||
headers.update(_build_auth_headers(matched_source))
|
headers.update(_build_auth_headers(matched_source))
|
||||||
auth = _get_basic_auth(matched_source) if matched_source else None
|
auth = _get_basic_auth(matched_source) if matched_source else None
|
||||||
|
|
||||||
timeout = httpx.Timeout(
|
timeout = httpx.Timeout(300.0, connect=PROXY_CONNECT_TIMEOUT) # 5 minutes for large files
|
||||||
connect=PROXY_CONNECT_TIMEOUT,
|
|
||||||
read=300.0, # 5 minutes for large files
|
|
||||||
)
|
|
||||||
|
|
||||||
# Fetch the file
|
# Fetch the file
|
||||||
logger.info(f"PyPI proxy: fetching {filename} from {upstream_url}")
|
logger.info(f"PyPI proxy: fetching {filename} from {upstream_url}")
|
||||||
@@ -436,20 +714,14 @@ async def pypi_download_file(
|
|||||||
content = response.content
|
content = response.content
|
||||||
content_type = response.headers.get('content-type', 'application/octet-stream')
|
content_type = response.headers.get('content-type', 'application/octet-stream')
|
||||||
|
|
||||||
# Compute hash
|
# Store in S3 (computes hash and deduplicates automatically)
|
||||||
sha256 = hashlib.sha256(content).hexdigest()
|
from io import BytesIO
|
||||||
size = len(content)
|
result = storage.store(BytesIO(content))
|
||||||
|
sha256 = result.sha256
|
||||||
|
size = result.size
|
||||||
|
|
||||||
logger.info(f"PyPI proxy: downloaded {filename}, {size} bytes, sha256={sha256[:12]}")
|
logger.info(f"PyPI proxy: downloaded {filename}, {size} bytes, sha256={sha256[:12]}")
|
||||||
|
|
||||||
# Store in S3
|
|
||||||
from io import BytesIO
|
|
||||||
artifact = storage.store_artifact(
|
|
||||||
file_obj=BytesIO(content),
|
|
||||||
filename=filename,
|
|
||||||
content_type=content_type,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Check if artifact already exists
|
# Check if artifact already exists
|
||||||
existing = db.query(Artifact).filter(Artifact.id == sha256).first()
|
existing = db.query(Artifact).filter(Artifact.id == sha256).first()
|
||||||
if existing:
|
if existing:
|
||||||
@@ -460,10 +732,15 @@ async def pypi_download_file(
|
|||||||
# Create artifact record
|
# Create artifact record
|
||||||
new_artifact = Artifact(
|
new_artifact = Artifact(
|
||||||
id=sha256,
|
id=sha256,
|
||||||
filename=filename,
|
original_name=filename,
|
||||||
content_type=content_type,
|
content_type=content_type,
|
||||||
size=size,
|
size=size,
|
||||||
ref_count=1,
|
ref_count=1,
|
||||||
|
created_by="pypi-proxy",
|
||||||
|
s3_key=result.s3_key,
|
||||||
|
checksum_md5=result.md5,
|
||||||
|
checksum_sha1=result.sha1,
|
||||||
|
s3_etag=result.s3_etag,
|
||||||
)
|
)
|
||||||
db.add(new_artifact)
|
db.add(new_artifact)
|
||||||
db.flush()
|
db.flush()
|
||||||
@@ -474,10 +751,16 @@ async def pypi_download_file(
|
|||||||
system_project = Project(
|
system_project = Project(
|
||||||
name="_pypi",
|
name="_pypi",
|
||||||
description="System project for cached PyPI packages",
|
description="System project for cached PyPI packages",
|
||||||
visibility="private",
|
is_public=True,
|
||||||
|
is_system=True,
|
||||||
|
created_by="pypi-proxy",
|
||||||
)
|
)
|
||||||
db.add(system_project)
|
db.add(system_project)
|
||||||
db.flush()
|
db.flush()
|
||||||
|
elif not system_project.is_system:
|
||||||
|
# Ensure existing project is marked as system
|
||||||
|
system_project.is_system = True
|
||||||
|
db.flush()
|
||||||
|
|
||||||
# Normalize package name
|
# Normalize package name
|
||||||
normalized_name = re.sub(r'[-_.]+', '-', package_name).lower()
|
normalized_name = re.sub(r'[-_.]+', '-', package_name).lower()
|
||||||
@@ -491,6 +774,7 @@ async def pypi_download_file(
|
|||||||
project_id=system_project.id,
|
project_id=system_project.id,
|
||||||
name=normalized_name,
|
name=normalized_name,
|
||||||
description=f"PyPI package: {normalized_name}",
|
description=f"PyPI package: {normalized_name}",
|
||||||
|
format="pypi",
|
||||||
)
|
)
|
||||||
db.add(package)
|
db.add(package)
|
||||||
db.flush()
|
db.flush()
|
||||||
@@ -505,9 +789,29 @@ async def pypi_download_file(
|
|||||||
package_id=package.id,
|
package_id=package.id,
|
||||||
name=filename,
|
name=filename,
|
||||||
artifact_id=sha256,
|
artifact_id=sha256,
|
||||||
|
created_by="pypi-proxy",
|
||||||
)
|
)
|
||||||
db.add(tag)
|
db.add(tag)
|
||||||
|
|
||||||
|
# Extract and create version
|
||||||
|
# Only create version for actual package files, not .metadata files
|
||||||
|
version = _extract_pypi_version(filename)
|
||||||
|
if version and not filename.endswith('.metadata'):
|
||||||
|
# Check by version string (the unique constraint is on package_id + version)
|
||||||
|
existing_version = db.query(PackageVersion).filter(
|
||||||
|
PackageVersion.package_id == package.id,
|
||||||
|
PackageVersion.version == version,
|
||||||
|
).first()
|
||||||
|
if not existing_version:
|
||||||
|
pkg_version = PackageVersion(
|
||||||
|
package_id=package.id,
|
||||||
|
artifact_id=sha256,
|
||||||
|
version=version,
|
||||||
|
version_source="filename",
|
||||||
|
created_by="pypi-proxy",
|
||||||
|
)
|
||||||
|
db.add(pkg_version)
|
||||||
|
|
||||||
# Cache the URL mapping
|
# Cache the URL mapping
|
||||||
existing_cached = db.query(CachedUrl).filter(CachedUrl.url_hash == url_hash).first()
|
existing_cached = db.query(CachedUrl).filter(CachedUrl.url_hash == url_hash).first()
|
||||||
if not existing_cached:
|
if not existing_cached:
|
||||||
@@ -518,8 +822,42 @@ async def pypi_download_file(
|
|||||||
)
|
)
|
||||||
db.add(cached_url_record)
|
db.add(cached_url_record)
|
||||||
|
|
||||||
|
# Extract and store dependencies
|
||||||
|
dependencies = _extract_dependencies(content, filename)
|
||||||
|
unique_deps = []
|
||||||
|
if dependencies:
|
||||||
|
# Deduplicate dependencies by package name (keep first occurrence)
|
||||||
|
seen_packages = set()
|
||||||
|
for dep_name, dep_version in dependencies:
|
||||||
|
if dep_name not in seen_packages:
|
||||||
|
seen_packages.add(dep_name)
|
||||||
|
unique_deps.append((dep_name, dep_version))
|
||||||
|
|
||||||
|
logger.info(f"PyPI proxy: extracted {len(unique_deps)} dependencies from {filename} (deduped from {len(dependencies)})")
|
||||||
|
for dep_name, dep_version in unique_deps:
|
||||||
|
# Check if this dependency already exists for this artifact
|
||||||
|
existing_dep = db.query(ArtifactDependency).filter(
|
||||||
|
ArtifactDependency.artifact_id == sha256,
|
||||||
|
ArtifactDependency.dependency_project == "_pypi",
|
||||||
|
ArtifactDependency.dependency_package == dep_name,
|
||||||
|
).first()
|
||||||
|
|
||||||
|
if not existing_dep:
|
||||||
|
dep = ArtifactDependency(
|
||||||
|
artifact_id=sha256,
|
||||||
|
dependency_project="_pypi",
|
||||||
|
dependency_package=dep_name,
|
||||||
|
version_constraint=dep_version if dep_version else "*",
|
||||||
|
)
|
||||||
|
db.add(dep)
|
||||||
|
|
||||||
db.commit()
|
db.commit()
|
||||||
|
|
||||||
|
# Proactively cache dependencies in the background
|
||||||
|
if unique_deps:
|
||||||
|
base_url = str(request.base_url).rstrip("/")
|
||||||
|
_start_background_dependency_caching(base_url, unique_deps)
|
||||||
|
|
||||||
# Return the file
|
# Return the file
|
||||||
return Response(
|
return Response(
|
||||||
content=content,
|
content=content,
|
||||||
|
|||||||
@@ -1680,6 +1680,7 @@ def create_project(
|
|||||||
name=db_project.name,
|
name=db_project.name,
|
||||||
description=db_project.description,
|
description=db_project.description,
|
||||||
is_public=db_project.is_public,
|
is_public=db_project.is_public,
|
||||||
|
is_system=db_project.is_system,
|
||||||
created_at=db_project.created_at,
|
created_at=db_project.created_at,
|
||||||
updated_at=db_project.updated_at,
|
updated_at=db_project.updated_at,
|
||||||
created_by=db_project.created_by,
|
created_by=db_project.created_by,
|
||||||
@@ -1704,6 +1705,7 @@ def get_project(
|
|||||||
name=project.name,
|
name=project.name,
|
||||||
description=project.description,
|
description=project.description,
|
||||||
is_public=project.is_public,
|
is_public=project.is_public,
|
||||||
|
is_system=project.is_system,
|
||||||
created_at=project.created_at,
|
created_at=project.created_at,
|
||||||
updated_at=project.updated_at,
|
updated_at=project.updated_at,
|
||||||
created_by=project.created_by,
|
created_by=project.created_by,
|
||||||
@@ -2704,6 +2706,7 @@ def list_team_projects(
|
|||||||
name=p.name,
|
name=p.name,
|
||||||
description=p.description,
|
description=p.description,
|
||||||
is_public=p.is_public,
|
is_public=p.is_public,
|
||||||
|
is_system=p.is_system,
|
||||||
created_at=p.created_at,
|
created_at=p.created_at,
|
||||||
updated_at=p.updated_at,
|
updated_at=p.updated_at,
|
||||||
created_by=p.created_by,
|
created_by=p.created_by,
|
||||||
@@ -2827,14 +2830,15 @@ def list_packages(
|
|||||||
db.query(func.count(Tag.id)).filter(Tag.package_id == pkg.id).scalar() or 0
|
db.query(func.count(Tag.id)).filter(Tag.package_id == pkg.id).scalar() or 0
|
||||||
)
|
)
|
||||||
|
|
||||||
# Get unique artifact count and total size via uploads
|
# Get unique artifact count and total size via tags
|
||||||
|
# (PyPI proxy creates tags without uploads, so query from tags)
|
||||||
artifact_stats = (
|
artifact_stats = (
|
||||||
db.query(
|
db.query(
|
||||||
func.count(func.distinct(Upload.artifact_id)),
|
func.count(func.distinct(Tag.artifact_id)),
|
||||||
func.coalesce(func.sum(Artifact.size), 0),
|
func.coalesce(func.sum(Artifact.size), 0),
|
||||||
)
|
)
|
||||||
.join(Artifact, Upload.artifact_id == Artifact.id)
|
.join(Artifact, Tag.artifact_id == Artifact.id)
|
||||||
.filter(Upload.package_id == pkg.id)
|
.filter(Tag.package_id == pkg.id)
|
||||||
.first()
|
.first()
|
||||||
)
|
)
|
||||||
artifact_count = artifact_stats[0] if artifact_stats else 0
|
artifact_count = artifact_stats[0] if artifact_stats else 0
|
||||||
@@ -2930,14 +2934,15 @@ def get_package(
|
|||||||
db.query(func.count(Tag.id)).filter(Tag.package_id == pkg.id).scalar() or 0
|
db.query(func.count(Tag.id)).filter(Tag.package_id == pkg.id).scalar() or 0
|
||||||
)
|
)
|
||||||
|
|
||||||
# Get unique artifact count and total size via uploads
|
# Get unique artifact count and total size via tags
|
||||||
|
# (PyPI proxy creates tags without uploads, so query from tags)
|
||||||
artifact_stats = (
|
artifact_stats = (
|
||||||
db.query(
|
db.query(
|
||||||
func.count(func.distinct(Upload.artifact_id)),
|
func.count(func.distinct(Tag.artifact_id)),
|
||||||
func.coalesce(func.sum(Artifact.size), 0),
|
func.coalesce(func.sum(Artifact.size), 0),
|
||||||
)
|
)
|
||||||
.join(Artifact, Upload.artifact_id == Artifact.id)
|
.join(Artifact, Tag.artifact_id == Artifact.id)
|
||||||
.filter(Upload.package_id == pkg.id)
|
.filter(Tag.package_id == pkg.id)
|
||||||
.first()
|
.first()
|
||||||
)
|
)
|
||||||
artifact_count = artifact_stats[0] if artifact_stats else 0
|
artifact_count = artifact_stats[0] if artifact_stats else 0
|
||||||
@@ -6280,14 +6285,14 @@ def get_package_stats(
|
|||||||
db.query(func.count(Tag.id)).filter(Tag.package_id == package.id).scalar() or 0
|
db.query(func.count(Tag.id)).filter(Tag.package_id == package.id).scalar() or 0
|
||||||
)
|
)
|
||||||
|
|
||||||
# Artifact stats via uploads
|
# Artifact stats via tags (tags exist for both user uploads and PyPI proxy)
|
||||||
artifact_stats = (
|
artifact_stats = (
|
||||||
db.query(
|
db.query(
|
||||||
func.count(func.distinct(Upload.artifact_id)),
|
func.count(func.distinct(Tag.artifact_id)),
|
||||||
func.coalesce(func.sum(Artifact.size), 0),
|
func.coalesce(func.sum(Artifact.size), 0),
|
||||||
)
|
)
|
||||||
.join(Artifact, Upload.artifact_id == Artifact.id)
|
.join(Artifact, Tag.artifact_id == Artifact.id)
|
||||||
.filter(Upload.package_id == package.id)
|
.filter(Tag.package_id == package.id)
|
||||||
.first()
|
.first()
|
||||||
)
|
)
|
||||||
artifact_count = artifact_stats[0] if artifact_stats else 0
|
artifact_count = artifact_stats[0] if artifact_stats else 0
|
||||||
|
|||||||
@@ -33,6 +33,7 @@ class ProjectResponse(BaseModel):
|
|||||||
name: str
|
name: str
|
||||||
description: Optional[str]
|
description: Optional[str]
|
||||||
is_public: bool
|
is_public: bool
|
||||||
|
is_system: bool = False
|
||||||
created_at: datetime
|
created_at: datetime
|
||||||
updated_at: datetime
|
updated_at: datetime
|
||||||
created_by: str
|
created_by: str
|
||||||
|
|||||||
1
backend/scripts/__init__.py
Normal file
1
backend/scripts/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
# Scripts package
|
||||||
262
backend/scripts/backfill_pypi_dependencies.py
Normal file
262
backend/scripts/backfill_pypi_dependencies.py
Normal file
@@ -0,0 +1,262 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
"""
|
||||||
|
Backfill script to extract dependencies from cached PyPI packages.
|
||||||
|
|
||||||
|
This script scans all artifacts in the _pypi project and extracts
|
||||||
|
Requires-Dist metadata from wheel and sdist files that don't already
|
||||||
|
have dependencies recorded.
|
||||||
|
|
||||||
|
Usage:
|
||||||
|
# From within the container:
|
||||||
|
python -m scripts.backfill_pypi_dependencies
|
||||||
|
|
||||||
|
# Or with docker exec:
|
||||||
|
docker exec orchard_orchard-server_1 python -m scripts.backfill_pypi_dependencies
|
||||||
|
|
||||||
|
# Dry run (preview only):
|
||||||
|
docker exec orchard_orchard-server_1 python -m scripts.backfill_pypi_dependencies --dry-run
|
||||||
|
"""
|
||||||
|
|
||||||
|
import argparse
|
||||||
|
import logging
|
||||||
|
import re
|
||||||
|
import sys
|
||||||
|
import tarfile
|
||||||
|
import zipfile
|
||||||
|
from io import BytesIO
|
||||||
|
from typing import List, Optional, Tuple
|
||||||
|
|
||||||
|
# Add parent directory to path for imports
|
||||||
|
sys.path.insert(0, "/app")
|
||||||
|
|
||||||
|
from sqlalchemy import create_engine
|
||||||
|
from sqlalchemy.orm import sessionmaker
|
||||||
|
|
||||||
|
from backend.app.config import get_settings
|
||||||
|
from backend.app.models import (
|
||||||
|
Artifact,
|
||||||
|
ArtifactDependency,
|
||||||
|
Package,
|
||||||
|
Project,
|
||||||
|
Tag,
|
||||||
|
)
|
||||||
|
from backend.app.storage import get_storage
|
||||||
|
|
||||||
|
logging.basicConfig(
|
||||||
|
level=logging.INFO,
|
||||||
|
format="%(asctime)s - %(levelname)s - %(message)s",
|
||||||
|
)
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def parse_requires_dist(requires_dist: str) -> Tuple[Optional[str], Optional[str]]:
|
||||||
|
"""Parse a Requires-Dist line into (package_name, version_constraint)."""
|
||||||
|
# Remove any environment markers (after semicolon)
|
||||||
|
if ";" in requires_dist:
|
||||||
|
requires_dist = requires_dist.split(";")[0].strip()
|
||||||
|
|
||||||
|
# Match patterns like "package (>=1.0)" or "package>=1.0" or "package"
|
||||||
|
match = re.match(
|
||||||
|
r"^([a-zA-Z0-9][-a-zA-Z0-9._]*)\s*(?:\(([^)]+)\)|([<>=!~][^\s;]+))?",
|
||||||
|
requires_dist.strip(),
|
||||||
|
)
|
||||||
|
|
||||||
|
if not match:
|
||||||
|
return None, None
|
||||||
|
|
||||||
|
package_name = match.group(1)
|
||||||
|
version_constraint = match.group(2) or match.group(3)
|
||||||
|
|
||||||
|
# Normalize package name (PEP 503)
|
||||||
|
normalized_name = re.sub(r"[-_.]+", "-", package_name).lower()
|
||||||
|
|
||||||
|
if version_constraint:
|
||||||
|
version_constraint = version_constraint.strip()
|
||||||
|
|
||||||
|
return normalized_name, version_constraint
|
||||||
|
|
||||||
|
|
||||||
|
def extract_requires_from_metadata(metadata_content: str) -> List[Tuple[str, Optional[str]]]:
|
||||||
|
"""Extract all Requires-Dist entries from METADATA/PKG-INFO content."""
|
||||||
|
dependencies = []
|
||||||
|
|
||||||
|
for line in metadata_content.split("\n"):
|
||||||
|
if line.startswith("Requires-Dist:"):
|
||||||
|
value = line[len("Requires-Dist:"):].strip()
|
||||||
|
pkg_name, version = parse_requires_dist(value)
|
||||||
|
if pkg_name:
|
||||||
|
dependencies.append((pkg_name, version))
|
||||||
|
|
||||||
|
return dependencies
|
||||||
|
|
||||||
|
|
||||||
|
def extract_metadata_from_wheel(content: bytes) -> Optional[str]:
|
||||||
|
"""Extract METADATA file content from a wheel (zip) file."""
|
||||||
|
try:
|
||||||
|
with zipfile.ZipFile(BytesIO(content)) as zf:
|
||||||
|
for name in zf.namelist():
|
||||||
|
if name.endswith(".dist-info/METADATA"):
|
||||||
|
return zf.read(name).decode("utf-8", errors="replace")
|
||||||
|
except Exception as e:
|
||||||
|
logger.warning(f"Failed to extract metadata from wheel: {e}")
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def extract_metadata_from_sdist(content: bytes) -> Optional[str]:
|
||||||
|
"""Extract PKG-INFO file content from a source distribution (.tar.gz)."""
|
||||||
|
try:
|
||||||
|
with tarfile.open(fileobj=BytesIO(content), mode="r:gz") as tf:
|
||||||
|
for member in tf.getmembers():
|
||||||
|
if member.name.endswith("/PKG-INFO") and member.name.count("/") == 1:
|
||||||
|
f = tf.extractfile(member)
|
||||||
|
if f:
|
||||||
|
return f.read().decode("utf-8", errors="replace")
|
||||||
|
except Exception as e:
|
||||||
|
logger.warning(f"Failed to extract metadata from sdist: {e}")
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def extract_dependencies(content: bytes, filename: str) -> List[Tuple[str, Optional[str]]]:
|
||||||
|
"""Extract dependencies from a PyPI package file."""
|
||||||
|
metadata = None
|
||||||
|
|
||||||
|
if filename.endswith(".whl"):
|
||||||
|
metadata = extract_metadata_from_wheel(content)
|
||||||
|
elif filename.endswith(".tar.gz"):
|
||||||
|
metadata = extract_metadata_from_sdist(content)
|
||||||
|
|
||||||
|
if metadata:
|
||||||
|
return extract_requires_from_metadata(metadata)
|
||||||
|
|
||||||
|
return []
|
||||||
|
|
||||||
|
|
||||||
|
def backfill_dependencies(dry_run: bool = False):
|
||||||
|
"""Main backfill function."""
|
||||||
|
settings = get_settings()
|
||||||
|
|
||||||
|
# Create database connection
|
||||||
|
engine = create_engine(settings.database_url)
|
||||||
|
Session = sessionmaker(bind=engine)
|
||||||
|
db = Session()
|
||||||
|
|
||||||
|
# Create storage client
|
||||||
|
storage = get_storage()
|
||||||
|
|
||||||
|
try:
|
||||||
|
# Find the _pypi project
|
||||||
|
pypi_project = db.query(Project).filter(Project.name == "_pypi").first()
|
||||||
|
if not pypi_project:
|
||||||
|
logger.info("No _pypi project found. Nothing to backfill.")
|
||||||
|
return
|
||||||
|
|
||||||
|
# Get all packages in _pypi
|
||||||
|
packages = db.query(Package).filter(Package.project_id == pypi_project.id).all()
|
||||||
|
logger.info(f"Found {len(packages)} packages in _pypi project")
|
||||||
|
|
||||||
|
total_artifacts = 0
|
||||||
|
artifacts_with_deps = 0
|
||||||
|
artifacts_processed = 0
|
||||||
|
dependencies_added = 0
|
||||||
|
|
||||||
|
for package in packages:
|
||||||
|
# Get all tags (each tag points to an artifact)
|
||||||
|
tags = db.query(Tag).filter(Tag.package_id == package.id).all()
|
||||||
|
|
||||||
|
for tag in tags:
|
||||||
|
total_artifacts += 1
|
||||||
|
filename = tag.name
|
||||||
|
|
||||||
|
# Skip non-package files (like .metadata files)
|
||||||
|
if not (filename.endswith(".whl") or filename.endswith(".tar.gz")):
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Check if this artifact already has dependencies
|
||||||
|
existing_deps = db.query(ArtifactDependency).filter(
|
||||||
|
ArtifactDependency.artifact_id == tag.artifact_id
|
||||||
|
).count()
|
||||||
|
|
||||||
|
if existing_deps > 0:
|
||||||
|
artifacts_with_deps += 1
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Get the artifact
|
||||||
|
artifact = db.query(Artifact).filter(Artifact.id == tag.artifact_id).first()
|
||||||
|
if not artifact:
|
||||||
|
logger.warning(f"Artifact {tag.artifact_id} not found for tag {filename}")
|
||||||
|
continue
|
||||||
|
|
||||||
|
logger.info(f"Processing {package.name}/{filename}...")
|
||||||
|
|
||||||
|
if dry_run:
|
||||||
|
logger.info(f" [DRY RUN] Would extract dependencies from {filename}")
|
||||||
|
artifacts_processed += 1
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Download the artifact from S3
|
||||||
|
try:
|
||||||
|
content = storage.get(artifact.s3_key)
|
||||||
|
except Exception as e:
|
||||||
|
logger.error(f" Failed to download {filename}: {e}")
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Extract dependencies
|
||||||
|
deps = extract_dependencies(content, filename)
|
||||||
|
|
||||||
|
if deps:
|
||||||
|
logger.info(f" Found {len(deps)} dependencies")
|
||||||
|
for dep_name, dep_version in deps:
|
||||||
|
# Check if already exists (race condition protection)
|
||||||
|
existing = db.query(ArtifactDependency).filter(
|
||||||
|
ArtifactDependency.artifact_id == tag.artifact_id,
|
||||||
|
ArtifactDependency.dependency_project == "_pypi",
|
||||||
|
ArtifactDependency.dependency_package == dep_name,
|
||||||
|
).first()
|
||||||
|
|
||||||
|
if not existing:
|
||||||
|
dep = ArtifactDependency(
|
||||||
|
artifact_id=tag.artifact_id,
|
||||||
|
dependency_project="_pypi",
|
||||||
|
dependency_package=dep_name,
|
||||||
|
version_constraint=dep_version if dep_version else "*",
|
||||||
|
)
|
||||||
|
db.add(dep)
|
||||||
|
dependencies_added += 1
|
||||||
|
logger.info(f" + {dep_name} {dep_version or '*'}")
|
||||||
|
|
||||||
|
db.commit()
|
||||||
|
else:
|
||||||
|
logger.info(f" No dependencies found")
|
||||||
|
|
||||||
|
artifacts_processed += 1
|
||||||
|
|
||||||
|
logger.info("")
|
||||||
|
logger.info("=" * 50)
|
||||||
|
logger.info("Backfill complete!")
|
||||||
|
logger.info(f" Total artifacts: {total_artifacts}")
|
||||||
|
logger.info(f" Already had deps: {artifacts_with_deps}")
|
||||||
|
logger.info(f" Processed: {artifacts_processed}")
|
||||||
|
logger.info(f" Dependencies added: {dependencies_added}")
|
||||||
|
if dry_run:
|
||||||
|
logger.info(" (DRY RUN - no changes made)")
|
||||||
|
|
||||||
|
finally:
|
||||||
|
db.close()
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
parser = argparse.ArgumentParser(
|
||||||
|
description="Backfill dependencies for cached PyPI packages"
|
||||||
|
)
|
||||||
|
parser.add_argument(
|
||||||
|
"--dry-run",
|
||||||
|
action="store_true",
|
||||||
|
help="Preview what would be done without making changes",
|
||||||
|
)
|
||||||
|
args = parser.parse_args()
|
||||||
|
|
||||||
|
backfill_dependencies(dry_run=args.dry_run)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
@@ -128,7 +128,9 @@ class TestProjectListingFilters:
|
|||||||
assert response.status_code == 200
|
assert response.status_code == 200
|
||||||
|
|
||||||
data = response.json()
|
data = response.json()
|
||||||
names = [p["name"] for p in data["items"]]
|
# Filter out system projects (names starting with "_") as they may have
|
||||||
|
# collation-specific sort behavior and aren't part of the test data
|
||||||
|
names = [p["name"] for p in data["items"] if not p["name"].startswith("_")]
|
||||||
assert names == sorted(names)
|
assert names == sorted(names)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -17,21 +17,31 @@ class TestPyPIProxyEndpoints:
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
@pytest.mark.integration
|
@pytest.mark.integration
|
||||||
def test_pypi_simple_index_no_sources(self):
|
def test_pypi_simple_index(self):
|
||||||
"""Test that /pypi/simple/ returns 503 when no sources configured."""
|
"""Test that /pypi/simple/ returns HTML response."""
|
||||||
with httpx.Client(base_url=get_base_url(), timeout=30.0) as client:
|
with httpx.Client(base_url=get_base_url(), timeout=30.0) as client:
|
||||||
response = client.get("/pypi/simple/")
|
response = client.get("/pypi/simple/")
|
||||||
# Should return 503 when no PyPI upstream sources are configured
|
# Returns 200 if sources configured, 503 if not
|
||||||
assert response.status_code == 503
|
assert response.status_code in (200, 503)
|
||||||
assert "No PyPI upstream sources configured" in response.json()["detail"]
|
if response.status_code == 200:
|
||||||
|
assert "text/html" in response.headers.get("content-type", "")
|
||||||
|
else:
|
||||||
|
assert "No PyPI upstream sources configured" in response.json()["detail"]
|
||||||
|
|
||||||
@pytest.mark.integration
|
@pytest.mark.integration
|
||||||
def test_pypi_package_no_sources(self):
|
def test_pypi_package_endpoint(self):
|
||||||
"""Test that /pypi/simple/{package}/ returns 503 when no sources configured."""
|
"""Test that /pypi/simple/{package}/ returns appropriate response."""
|
||||||
with httpx.Client(base_url=get_base_url(), timeout=30.0) as client:
|
with httpx.Client(base_url=get_base_url(), timeout=30.0) as client:
|
||||||
response = client.get("/pypi/simple/requests/")
|
response = client.get("/pypi/simple/requests/")
|
||||||
assert response.status_code == 503
|
# Returns 200 if sources configured and package found,
|
||||||
assert "No PyPI upstream sources configured" in response.json()["detail"]
|
# 404 if package not found, 503 if no sources
|
||||||
|
assert response.status_code in (200, 404, 503)
|
||||||
|
if response.status_code == 200:
|
||||||
|
assert "text/html" in response.headers.get("content-type", "")
|
||||||
|
elif response.status_code == 404:
|
||||||
|
assert "not found" in response.json()["detail"].lower()
|
||||||
|
else: # 503
|
||||||
|
assert "No PyPI upstream sources configured" in response.json()["detail"]
|
||||||
|
|
||||||
@pytest.mark.integration
|
@pytest.mark.integration
|
||||||
def test_pypi_download_missing_upstream_param(self):
|
def test_pypi_download_missing_upstream_param(self):
|
||||||
@@ -58,7 +68,13 @@ class TestPyPILinkRewriting:
|
|||||||
</html>
|
</html>
|
||||||
'''
|
'''
|
||||||
|
|
||||||
result = _rewrite_package_links(html, "http://localhost:8080", "requests")
|
# upstream_base_url is used to resolve relative URLs (not needed here since URLs are absolute)
|
||||||
|
result = _rewrite_package_links(
|
||||||
|
html,
|
||||||
|
"http://localhost:8080",
|
||||||
|
"requests",
|
||||||
|
"https://pypi.org/simple/requests/"
|
||||||
|
)
|
||||||
|
|
||||||
# Links should be rewritten to go through our proxy
|
# Links should be rewritten to go through our proxy
|
||||||
assert "/pypi/simple/requests/requests-2.31.0.tar.gz?upstream=" in result
|
assert "/pypi/simple/requests/requests-2.31.0.tar.gz?upstream=" in result
|
||||||
@@ -69,25 +85,53 @@ class TestPyPILinkRewriting:
|
|||||||
assert "#sha256=abc123" in result
|
assert "#sha256=abc123" in result
|
||||||
assert "#sha256=def456" in result
|
assert "#sha256=def456" in result
|
||||||
|
|
||||||
|
def test_rewrite_relative_links(self):
|
||||||
|
"""Test that relative URLs are resolved to absolute URLs."""
|
||||||
|
from app.pypi_proxy import _rewrite_package_links
|
||||||
|
|
||||||
|
# Artifactory-style relative URLs
|
||||||
|
html = '''
|
||||||
|
<html>
|
||||||
|
<body>
|
||||||
|
<a href="../../packages/ab/cd/requests-2.31.0.tar.gz#sha256=abc123">requests-2.31.0.tar.gz</a>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
|
'''
|
||||||
|
|
||||||
|
result = _rewrite_package_links(
|
||||||
|
html,
|
||||||
|
"https://orchard.example.com",
|
||||||
|
"requests",
|
||||||
|
"https://artifactory.example.com/api/pypi/pypi-remote/simple/requests/"
|
||||||
|
)
|
||||||
|
|
||||||
|
# The relative URL should be resolved to absolute
|
||||||
|
# ../../packages/ab/cd/... from /api/pypi/pypi-remote/simple/requests/ resolves to /api/pypi/pypi-remote/packages/ab/cd/...
|
||||||
|
assert "upstream=https%3A%2F%2Fartifactory.example.com%2Fapi%2Fpypi%2Fpypi-remote%2Fpackages" in result
|
||||||
|
# Hash fragment should be preserved
|
||||||
|
assert "#sha256=abc123" in result
|
||||||
|
|
||||||
|
|
||||||
class TestPyPIPackageNormalization:
|
class TestPyPIPackageNormalization:
|
||||||
"""Tests for PyPI package name normalization."""
|
"""Tests for PyPI package name normalization."""
|
||||||
|
|
||||||
@pytest.mark.integration
|
@pytest.mark.integration
|
||||||
def test_package_name_normalized(self):
|
def test_package_name_normalized(self):
|
||||||
"""Test that package names are normalized per PEP 503."""
|
"""Test that package names are normalized per PEP 503.
|
||||||
# These should all be treated the same:
|
|
||||||
# requests, Requests, requests_, requests-
|
|
||||||
# The endpoint normalizes to lowercase with hyphens
|
|
||||||
|
|
||||||
|
Different capitalizations/separators should all be valid paths.
|
||||||
|
The endpoint normalizes to lowercase with hyphens before lookup.
|
||||||
|
"""
|
||||||
with httpx.Client(base_url=get_base_url(), timeout=30.0) as client:
|
with httpx.Client(base_url=get_base_url(), timeout=30.0) as client:
|
||||||
# Without upstream sources, we get 503, but the normalization
|
# Test various name formats - all should be valid endpoint paths
|
||||||
# happens before the source lookup
|
for package_name in ["Requests", "some_package", "some-package"]:
|
||||||
response = client.get("/pypi/simple/Requests/")
|
response = client.get(f"/pypi/simple/{package_name}/")
|
||||||
assert response.status_code == 503 # No sources, but path was valid
|
# 200 = found, 404 = not found, 503 = no sources configured
|
||||||
|
assert response.status_code in (200, 404, 503), \
|
||||||
|
f"Unexpected status {response.status_code} for {package_name}"
|
||||||
|
|
||||||
response = client.get("/pypi/simple/some_package/")
|
# Verify response is appropriate for the status code
|
||||||
assert response.status_code == 503
|
if response.status_code == 200:
|
||||||
|
assert "text/html" in response.headers.get("content-type", "")
|
||||||
response = client.get("/pypi/simple/some-package/")
|
elif response.status_code == 503:
|
||||||
assert response.status_code == 503
|
assert "No PyPI upstream sources configured" in response.json()["detail"]
|
||||||
|
|||||||
@@ -132,6 +132,12 @@
|
|||||||
color: #c62828;
|
color: #c62828;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.coming-soon-badge {
|
||||||
|
color: #9e9e9e;
|
||||||
|
font-style: italic;
|
||||||
|
font-size: 0.85em;
|
||||||
|
}
|
||||||
|
|
||||||
/* Actions */
|
/* Actions */
|
||||||
.actions-cell {
|
.actions-cell {
|
||||||
white-space: nowrap;
|
white-space: nowrap;
|
||||||
|
|||||||
@@ -12,6 +12,7 @@ import { UpstreamSource, SourceType, AuthType } from '../types';
|
|||||||
import './AdminCachePage.css';
|
import './AdminCachePage.css';
|
||||||
|
|
||||||
const SOURCE_TYPES: SourceType[] = ['npm', 'pypi', 'maven', 'docker', 'helm', 'nuget', 'deb', 'rpm', 'generic'];
|
const SOURCE_TYPES: SourceType[] = ['npm', 'pypi', 'maven', 'docker', 'helm', 'nuget', 'deb', 'rpm', 'generic'];
|
||||||
|
const SUPPORTED_SOURCE_TYPES: Set<SourceType> = new Set(['pypi', 'generic']);
|
||||||
const AUTH_TYPES: AuthType[] = ['none', 'basic', 'bearer', 'api_key'];
|
const AUTH_TYPES: AuthType[] = ['none', 'basic', 'bearer', 'api_key'];
|
||||||
|
|
||||||
function AdminCachePage() {
|
function AdminCachePage() {
|
||||||
@@ -285,7 +286,12 @@ function AdminCachePage() {
|
|||||||
<span className="env-badge" title="Defined via environment variable">ENV</span>
|
<span className="env-badge" title="Defined via environment variable">ENV</span>
|
||||||
)}
|
)}
|
||||||
</td>
|
</td>
|
||||||
<td>{source.source_type}</td>
|
<td>
|
||||||
|
{source.source_type}
|
||||||
|
{!SUPPORTED_SOURCE_TYPES.has(source.source_type) && (
|
||||||
|
<span className="coming-soon-badge"> (coming soon)</span>
|
||||||
|
)}
|
||||||
|
</td>
|
||||||
<td className="url-cell" title={source.url}>{source.url}</td>
|
<td className="url-cell" title={source.url}>{source.url}</td>
|
||||||
<td>{source.priority}</td>
|
<td>{source.priority}</td>
|
||||||
<td>
|
<td>
|
||||||
@@ -359,7 +365,7 @@ function AdminCachePage() {
|
|||||||
>
|
>
|
||||||
{SOURCE_TYPES.map((type) => (
|
{SOURCE_TYPES.map((type) => (
|
||||||
<option key={type} value={type}>
|
<option key={type} value={type}>
|
||||||
{type}
|
{type}{!SUPPORTED_SOURCE_TYPES.has(type) ? ' (coming soon)' : ''}
|
||||||
</option>
|
</option>
|
||||||
))}
|
))}
|
||||||
</select>
|
</select>
|
||||||
|
|||||||
@@ -249,7 +249,7 @@ function Home() {
|
|||||||
key: 'created_by',
|
key: 'created_by',
|
||||||
header: 'Owner',
|
header: 'Owner',
|
||||||
className: 'cell-owner',
|
className: 'cell-owner',
|
||||||
render: (project) => project.created_by,
|
render: (project) => project.team_name || project.created_by,
|
||||||
},
|
},
|
||||||
...(user
|
...(user
|
||||||
? [
|
? [
|
||||||
|
|||||||
@@ -642,6 +642,11 @@ tr:hover .copy-btn {
|
|||||||
padding: 20px;
|
padding: 20px;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Ensure file modal needs higher z-index when opened from deps modal */
|
||||||
|
.modal-overlay:has(.ensure-file-modal) {
|
||||||
|
z-index: 1100;
|
||||||
|
}
|
||||||
|
|
||||||
.ensure-file-modal {
|
.ensure-file-modal {
|
||||||
background: var(--bg-secondary);
|
background: var(--bg-secondary);
|
||||||
border: 1px solid var(--border-primary);
|
border: 1px solid var(--border-primary);
|
||||||
@@ -793,4 +798,194 @@ tr:hover .copy-btn {
|
|||||||
.ensure-file-modal {
|
.ensure-file-modal {
|
||||||
max-height: 90vh;
|
max-height: 90vh;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.action-menu-dropdown {
|
||||||
|
right: 0;
|
||||||
|
left: auto;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Header upload button */
|
||||||
|
.header-upload-btn {
|
||||||
|
margin-left: auto;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Tag/Version cell */
|
||||||
|
.tag-version-cell {
|
||||||
|
display: flex;
|
||||||
|
flex-direction: column;
|
||||||
|
gap: 4px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.tag-version-cell .version-badge {
|
||||||
|
font-size: 0.75rem;
|
||||||
|
color: var(--text-muted);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Icon buttons */
|
||||||
|
.btn-icon {
|
||||||
|
display: flex;
|
||||||
|
align-items: center;
|
||||||
|
justify-content: center;
|
||||||
|
width: 32px;
|
||||||
|
height: 32px;
|
||||||
|
padding: 0;
|
||||||
|
background: transparent;
|
||||||
|
border: 1px solid transparent;
|
||||||
|
border-radius: var(--radius-sm);
|
||||||
|
color: var(--text-secondary);
|
||||||
|
cursor: pointer;
|
||||||
|
transition: all var(--transition-fast);
|
||||||
|
}
|
||||||
|
|
||||||
|
.btn-icon:hover {
|
||||||
|
background: var(--bg-hover);
|
||||||
|
color: var(--text-primary);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Action menu */
|
||||||
|
.action-buttons {
|
||||||
|
display: flex;
|
||||||
|
align-items: center;
|
||||||
|
gap: 4px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.action-menu {
|
||||||
|
position: relative;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Action menu backdrop for click-outside */
|
||||||
|
.action-menu-backdrop {
|
||||||
|
position: fixed;
|
||||||
|
top: 0;
|
||||||
|
left: 0;
|
||||||
|
right: 0;
|
||||||
|
bottom: 0;
|
||||||
|
z-index: 999;
|
||||||
|
}
|
||||||
|
|
||||||
|
.action-menu-dropdown {
|
||||||
|
position: fixed;
|
||||||
|
z-index: 1000;
|
||||||
|
min-width: 180px;
|
||||||
|
padding: 4px 0;
|
||||||
|
background: var(--bg-secondary);
|
||||||
|
border: 1px solid var(--border-primary);
|
||||||
|
border-radius: var(--radius-md);
|
||||||
|
box-shadow: 0 4px 12px rgba(0, 0, 0, 0.15);
|
||||||
|
}
|
||||||
|
|
||||||
|
.action-menu-dropdown button {
|
||||||
|
display: block;
|
||||||
|
width: 100%;
|
||||||
|
padding: 8px 12px;
|
||||||
|
background: none;
|
||||||
|
border: none;
|
||||||
|
text-align: left;
|
||||||
|
font-size: 0.875rem;
|
||||||
|
color: var(--text-primary);
|
||||||
|
cursor: pointer;
|
||||||
|
transition: background var(--transition-fast);
|
||||||
|
}
|
||||||
|
|
||||||
|
.action-menu-dropdown button:hover {
|
||||||
|
background: var(--bg-hover);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Upload Modal */
|
||||||
|
.upload-modal,
|
||||||
|
.create-tag-modal {
|
||||||
|
background: var(--bg-secondary);
|
||||||
|
border-radius: var(--radius-lg);
|
||||||
|
width: 90%;
|
||||||
|
max-width: 500px;
|
||||||
|
max-height: 90vh;
|
||||||
|
overflow: hidden;
|
||||||
|
}
|
||||||
|
|
||||||
|
.modal-header {
|
||||||
|
display: flex;
|
||||||
|
align-items: center;
|
||||||
|
justify-content: space-between;
|
||||||
|
padding: 16px 20px;
|
||||||
|
border-bottom: 1px solid var(--border-primary);
|
||||||
|
}
|
||||||
|
|
||||||
|
.modal-header h3 {
|
||||||
|
margin: 0;
|
||||||
|
font-size: 1.125rem;
|
||||||
|
font-weight: 600;
|
||||||
|
}
|
||||||
|
|
||||||
|
.modal-body {
|
||||||
|
padding: 20px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.modal-description {
|
||||||
|
margin-bottom: 16px;
|
||||||
|
color: var(--text-secondary);
|
||||||
|
font-size: 0.875rem;
|
||||||
|
}
|
||||||
|
|
||||||
|
.modal-actions {
|
||||||
|
display: flex;
|
||||||
|
justify-content: flex-end;
|
||||||
|
gap: 12px;
|
||||||
|
margin-top: 20px;
|
||||||
|
padding-top: 16px;
|
||||||
|
border-top: 1px solid var(--border-primary);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Dependencies Modal */
|
||||||
|
.deps-modal {
|
||||||
|
background: var(--bg-secondary);
|
||||||
|
border-radius: var(--radius-lg);
|
||||||
|
width: 90%;
|
||||||
|
max-width: 600px;
|
||||||
|
max-height: 80vh;
|
||||||
|
overflow: hidden;
|
||||||
|
display: flex;
|
||||||
|
flex-direction: column;
|
||||||
|
}
|
||||||
|
|
||||||
|
.deps-modal .modal-body {
|
||||||
|
overflow-y: auto;
|
||||||
|
flex: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
.deps-modal-controls {
|
||||||
|
display: flex;
|
||||||
|
gap: 8px;
|
||||||
|
margin-bottom: 16px;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Artifact ID Modal */
|
||||||
|
.artifact-id-modal {
|
||||||
|
background: var(--bg-secondary);
|
||||||
|
border-radius: var(--radius-lg);
|
||||||
|
width: 90%;
|
||||||
|
max-width: 500px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.artifact-id-display {
|
||||||
|
display: flex;
|
||||||
|
align-items: center;
|
||||||
|
gap: 12px;
|
||||||
|
padding: 16px;
|
||||||
|
background: var(--bg-tertiary);
|
||||||
|
border-radius: var(--radius-md);
|
||||||
|
border: 1px solid var(--border-primary);
|
||||||
|
}
|
||||||
|
|
||||||
|
.artifact-id-display code {
|
||||||
|
font-family: 'JetBrains Mono', 'Fira Code', 'Consolas', monospace;
|
||||||
|
font-size: 0.8125rem;
|
||||||
|
color: var(--text-primary);
|
||||||
|
word-break: break-all;
|
||||||
|
flex: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
.artifact-id-display .copy-btn {
|
||||||
|
opacity: 1;
|
||||||
|
flex-shrink: 0;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -63,12 +63,17 @@ function PackagePage() {
|
|||||||
const [accessDenied, setAccessDenied] = useState(false);
|
const [accessDenied, setAccessDenied] = useState(false);
|
||||||
const [uploadTag, setUploadTag] = useState('');
|
const [uploadTag, setUploadTag] = useState('');
|
||||||
const [uploadSuccess, setUploadSuccess] = useState<string | null>(null);
|
const [uploadSuccess, setUploadSuccess] = useState<string | null>(null);
|
||||||
const [artifactIdInput, setArtifactIdInput] = useState('');
|
|
||||||
const [accessLevel, setAccessLevel] = useState<AccessLevel | null>(null);
|
const [accessLevel, setAccessLevel] = useState<AccessLevel | null>(null);
|
||||||
const [createTagName, setCreateTagName] = useState('');
|
const [createTagName, setCreateTagName] = useState('');
|
||||||
const [createTagArtifactId, setCreateTagArtifactId] = useState('');
|
const [createTagArtifactId, setCreateTagArtifactId] = useState('');
|
||||||
const [createTagLoading, setCreateTagLoading] = useState(false);
|
const [createTagLoading, setCreateTagLoading] = useState(false);
|
||||||
|
|
||||||
|
// UI state
|
||||||
|
const [showUploadModal, setShowUploadModal] = useState(false);
|
||||||
|
const [showCreateTagModal, setShowCreateTagModal] = useState(false);
|
||||||
|
const [openMenuId, setOpenMenuId] = useState<string | null>(null);
|
||||||
|
const [menuPosition, setMenuPosition] = useState<{ top: number; left: number } | null>(null);
|
||||||
|
|
||||||
// Dependencies state
|
// Dependencies state
|
||||||
const [selectedTag, setSelectedTag] = useState<TagDetail | null>(null);
|
const [selectedTag, setSelectedTag] = useState<TagDetail | null>(null);
|
||||||
const [dependencies, setDependencies] = useState<Dependency[]>([]);
|
const [dependencies, setDependencies] = useState<Dependency[]>([]);
|
||||||
@@ -86,6 +91,13 @@ function PackagePage() {
|
|||||||
// Dependency graph modal state
|
// Dependency graph modal state
|
||||||
const [showGraph, setShowGraph] = useState(false);
|
const [showGraph, setShowGraph] = useState(false);
|
||||||
|
|
||||||
|
// Dependencies modal state
|
||||||
|
const [showDepsModal, setShowDepsModal] = useState(false);
|
||||||
|
|
||||||
|
// Artifact ID modal state
|
||||||
|
const [showArtifactIdModal, setShowArtifactIdModal] = useState(false);
|
||||||
|
const [viewArtifactId, setViewArtifactId] = useState<string | null>(null);
|
||||||
|
|
||||||
// Ensure file modal state
|
// Ensure file modal state
|
||||||
const [showEnsureFile, setShowEnsureFile] = useState(false);
|
const [showEnsureFile, setShowEnsureFile] = useState(false);
|
||||||
const [ensureFileContent, setEnsureFileContent] = useState<string | null>(null);
|
const [ensureFileContent, setEnsureFileContent] = useState<string | null>(null);
|
||||||
@@ -96,6 +108,9 @@ function PackagePage() {
|
|||||||
// Derived permissions
|
// Derived permissions
|
||||||
const canWrite = accessLevel === 'write' || accessLevel === 'admin';
|
const canWrite = accessLevel === 'write' || accessLevel === 'admin';
|
||||||
|
|
||||||
|
// Detect system projects (convention: name starts with "_")
|
||||||
|
const isSystemProject = projectName?.startsWith('_') ?? false;
|
||||||
|
|
||||||
// Get params from URL
|
// Get params from URL
|
||||||
const page = parseInt(searchParams.get('page') || '1', 10);
|
const page = parseInt(searchParams.get('page') || '1', 10);
|
||||||
const search = searchParams.get('search') || '';
|
const search = searchParams.get('search') || '';
|
||||||
@@ -323,92 +338,212 @@ function PackagePage() {
|
|||||||
setSelectedTag(tag);
|
setSelectedTag(tag);
|
||||||
};
|
};
|
||||||
|
|
||||||
const columns = [
|
const handleMenuOpen = (e: React.MouseEvent, tagId: string) => {
|
||||||
{
|
e.stopPropagation();
|
||||||
key: 'name',
|
if (openMenuId === tagId) {
|
||||||
header: 'Tag',
|
setOpenMenuId(null);
|
||||||
sortable: true,
|
setMenuPosition(null);
|
||||||
render: (t: TagDetail) => (
|
} else {
|
||||||
<strong
|
const rect = e.currentTarget.getBoundingClientRect();
|
||||||
className={`tag-name-link ${selectedTag?.id === t.id ? 'selected' : ''}`}
|
setMenuPosition({ top: rect.bottom + 4, left: rect.right - 180 });
|
||||||
onClick={() => handleTagSelect(t)}
|
setOpenMenuId(tagId);
|
||||||
style={{ cursor: 'pointer' }}
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
// System projects show Version first, regular projects show Tag first
|
||||||
|
const columns = isSystemProject
|
||||||
|
? [
|
||||||
|
// System project columns: Version first, then Filename
|
||||||
|
{
|
||||||
|
key: 'version',
|
||||||
|
header: 'Version',
|
||||||
|
sortable: true,
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<strong
|
||||||
|
className={`tag-name-link ${selectedTag?.id === t.id ? 'selected' : ''}`}
|
||||||
|
onClick={() => handleTagSelect(t)}
|
||||||
|
style={{ cursor: 'pointer' }}
|
||||||
|
>
|
||||||
|
<span className="version-badge">{t.version || t.name}</span>
|
||||||
|
</strong>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'artifact_original_name',
|
||||||
|
header: 'Filename',
|
||||||
|
className: 'cell-truncate',
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<span title={t.artifact_original_name || t.name}>{t.artifact_original_name || t.name}</span>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'artifact_size',
|
||||||
|
header: 'Size',
|
||||||
|
render: (t: TagDetail) => <span>{formatBytes(t.artifact_size)}</span>,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'created_at',
|
||||||
|
header: 'Cached',
|
||||||
|
sortable: true,
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<span>{new Date(t.created_at).toLocaleDateString()}</span>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'actions',
|
||||||
|
header: '',
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<div className="action-buttons">
|
||||||
|
<a
|
||||||
|
href={getDownloadUrl(projectName!, packageName!, t.name)}
|
||||||
|
className="btn btn-icon"
|
||||||
|
download
|
||||||
|
title="Download"
|
||||||
|
>
|
||||||
|
<svg width="16" height="16" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2">
|
||||||
|
<path d="M21 15v4a2 2 0 0 1-2 2H5a2 2 0 0 1-2-2v-4" />
|
||||||
|
<polyline points="7 10 12 15 17 10" />
|
||||||
|
<line x1="12" y1="15" x2="12" y2="3" />
|
||||||
|
</svg>
|
||||||
|
</a>
|
||||||
|
<button
|
||||||
|
className="btn btn-icon"
|
||||||
|
onClick={(e) => handleMenuOpen(e, t.id)}
|
||||||
|
title="More actions"
|
||||||
|
>
|
||||||
|
<svg width="16" height="16" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2">
|
||||||
|
<circle cx="12" cy="12" r="1" />
|
||||||
|
<circle cx="12" cy="5" r="1" />
|
||||||
|
<circle cx="12" cy="19" r="1" />
|
||||||
|
</svg>
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
]
|
||||||
|
: [
|
||||||
|
// Regular project columns: Tag, Version, Filename
|
||||||
|
{
|
||||||
|
key: 'name',
|
||||||
|
header: 'Tag',
|
||||||
|
sortable: true,
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<strong
|
||||||
|
className={`tag-name-link ${selectedTag?.id === t.id ? 'selected' : ''}`}
|
||||||
|
onClick={() => handleTagSelect(t)}
|
||||||
|
style={{ cursor: 'pointer' }}
|
||||||
|
>
|
||||||
|
{t.name}
|
||||||
|
</strong>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'version',
|
||||||
|
header: 'Version',
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<span className="version-badge">{t.version || '—'}</span>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'artifact_original_name',
|
||||||
|
header: 'Filename',
|
||||||
|
className: 'cell-truncate',
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<span title={t.artifact_original_name || undefined}>{t.artifact_original_name || '—'}</span>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'artifact_size',
|
||||||
|
header: 'Size',
|
||||||
|
render: (t: TagDetail) => <span>{formatBytes(t.artifact_size)}</span>,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'created_at',
|
||||||
|
header: 'Created',
|
||||||
|
sortable: true,
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<span title={`by ${t.created_by}`}>{new Date(t.created_at).toLocaleDateString()}</span>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'actions',
|
||||||
|
header: '',
|
||||||
|
render: (t: TagDetail) => (
|
||||||
|
<div className="action-buttons">
|
||||||
|
<a
|
||||||
|
href={getDownloadUrl(projectName!, packageName!, t.name)}
|
||||||
|
className="btn btn-icon"
|
||||||
|
download
|
||||||
|
title="Download"
|
||||||
|
>
|
||||||
|
<svg width="16" height="16" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2">
|
||||||
|
<path d="M21 15v4a2 2 0 0 1-2 2H5a2 2 0 0 1-2-2v-4" />
|
||||||
|
<polyline points="7 10 12 15 17 10" />
|
||||||
|
<line x1="12" y1="15" x2="12" y2="3" />
|
||||||
|
</svg>
|
||||||
|
</a>
|
||||||
|
<button
|
||||||
|
className="btn btn-icon"
|
||||||
|
onClick={(e) => handleMenuOpen(e, t.id)}
|
||||||
|
title="More actions"
|
||||||
|
>
|
||||||
|
<svg width="16" height="16" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2">
|
||||||
|
<circle cx="12" cy="12" r="1" />
|
||||||
|
<circle cx="12" cy="5" r="1" />
|
||||||
|
<circle cx="12" cy="19" r="1" />
|
||||||
|
</svg>
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
// Find the tag for the open menu
|
||||||
|
const openMenuTag = tags.find(t => t.id === openMenuId);
|
||||||
|
|
||||||
|
// Close menu when clicking outside
|
||||||
|
const handleClickOutside = () => {
|
||||||
|
if (openMenuId) {
|
||||||
|
setOpenMenuId(null);
|
||||||
|
setMenuPosition(null);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
// Render dropdown menu as a portal-like element
|
||||||
|
const renderActionMenu = () => {
|
||||||
|
if (!openMenuId || !menuPosition || !openMenuTag) return null;
|
||||||
|
const t = openMenuTag;
|
||||||
|
return (
|
||||||
|
<div
|
||||||
|
className="action-menu-backdrop"
|
||||||
|
onClick={handleClickOutside}
|
||||||
|
>
|
||||||
|
<div
|
||||||
|
className="action-menu-dropdown"
|
||||||
|
style={{ top: menuPosition.top, left: menuPosition.left }}
|
||||||
|
onClick={(e) => e.stopPropagation()}
|
||||||
>
|
>
|
||||||
{t.name}
|
<button onClick={() => { setViewArtifactId(t.artifact_id); setShowArtifactIdModal(true); setOpenMenuId(null); setMenuPosition(null); }}>
|
||||||
</strong>
|
View Artifact ID
|
||||||
),
|
</button>
|
||||||
},
|
<button onClick={() => { navigator.clipboard.writeText(t.artifact_id); setOpenMenuId(null); setMenuPosition(null); }}>
|
||||||
{
|
Copy Artifact ID
|
||||||
key: 'version',
|
</button>
|
||||||
header: 'Version',
|
<button onClick={() => { fetchEnsureFileForTag(t.name); setOpenMenuId(null); setMenuPosition(null); }}>
|
||||||
render: (t: TagDetail) => (
|
View Ensure File
|
||||||
<span className="version-badge">{t.version || '-'}</span>
|
</button>
|
||||||
),
|
{canWrite && !isSystemProject && (
|
||||||
},
|
<button onClick={() => { setCreateTagArtifactId(t.artifact_id); setShowCreateTagModal(true); setOpenMenuId(null); setMenuPosition(null); }}>
|
||||||
{
|
Create/Update Tag
|
||||||
key: 'artifact_id',
|
</button>
|
||||||
header: 'Artifact ID',
|
)}
|
||||||
render: (t: TagDetail) => (
|
<button onClick={() => { handleTagSelect(t); setShowDepsModal(true); setOpenMenuId(null); setMenuPosition(null); }}>
|
||||||
<div className="artifact-id-cell">
|
View Dependencies
|
||||||
<code className="artifact-id">{t.artifact_id.substring(0, 12)}...</code>
|
|
||||||
<CopyButton text={t.artifact_id} />
|
|
||||||
</div>
|
|
||||||
),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'artifact_size',
|
|
||||||
header: 'Size',
|
|
||||||
render: (t: TagDetail) => <span>{formatBytes(t.artifact_size)}</span>,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'artifact_content_type',
|
|
||||||
header: 'Type',
|
|
||||||
render: (t: TagDetail) => (
|
|
||||||
<span className="content-type">{t.artifact_content_type || '-'}</span>
|
|
||||||
),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'artifact_original_name',
|
|
||||||
header: 'Filename',
|
|
||||||
className: 'cell-truncate',
|
|
||||||
render: (t: TagDetail) => (
|
|
||||||
<span title={t.artifact_original_name || undefined}>{t.artifact_original_name || '-'}</span>
|
|
||||||
),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'created_at',
|
|
||||||
header: 'Created',
|
|
||||||
sortable: true,
|
|
||||||
render: (t: TagDetail) => (
|
|
||||||
<div className="created-cell">
|
|
||||||
<span>{new Date(t.created_at).toLocaleString()}</span>
|
|
||||||
<span className="created-by">by {t.created_by}</span>
|
|
||||||
</div>
|
|
||||||
),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'actions',
|
|
||||||
header: 'Actions',
|
|
||||||
render: (t: TagDetail) => (
|
|
||||||
<div className="action-buttons">
|
|
||||||
<button
|
|
||||||
className="btn btn-secondary btn-small"
|
|
||||||
onClick={() => fetchEnsureFileForTag(t.name)}
|
|
||||||
title="View orchard.ensure file"
|
|
||||||
>
|
|
||||||
Ensure
|
|
||||||
</button>
|
</button>
|
||||||
<a
|
|
||||||
href={getDownloadUrl(projectName!, packageName!, t.name)}
|
|
||||||
className="btn btn-secondary btn-small"
|
|
||||||
download
|
|
||||||
>
|
|
||||||
Download
|
|
||||||
</a>
|
|
||||||
</div>
|
</div>
|
||||||
),
|
</div>
|
||||||
},
|
);
|
||||||
];
|
};
|
||||||
|
|
||||||
if (loading && !tagsData) {
|
if (loading && !tagsData) {
|
||||||
return <div className="loading">Loading...</div>;
|
return <div className="loading">Loading...</div>;
|
||||||
@@ -451,6 +586,19 @@ function PackagePage() {
|
|||||||
<div className="page-header__title-row">
|
<div className="page-header__title-row">
|
||||||
<h1>{packageName}</h1>
|
<h1>{packageName}</h1>
|
||||||
{pkg && <Badge variant="default">{pkg.format}</Badge>}
|
{pkg && <Badge variant="default">{pkg.format}</Badge>}
|
||||||
|
{user && canWrite && !isSystemProject && (
|
||||||
|
<button
|
||||||
|
className="btn btn-primary btn-small header-upload-btn"
|
||||||
|
onClick={() => setShowUploadModal(true)}
|
||||||
|
>
|
||||||
|
<svg width="14" height="14" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2" style={{ marginRight: '6px' }}>
|
||||||
|
<path d="M21 15v4a2 2 0 0 1-2 2H5a2 2 0 0 1-2-2v-4" />
|
||||||
|
<polyline points="17 8 12 3 7 8" />
|
||||||
|
<line x1="12" y1="3" x2="12" y2="15" />
|
||||||
|
</svg>
|
||||||
|
Upload
|
||||||
|
</button>
|
||||||
|
)}
|
||||||
</div>
|
</div>
|
||||||
{pkg?.description && <p className="description">{pkg.description}</p>}
|
{pkg?.description && <p className="description">{pkg.description}</p>}
|
||||||
<div className="page-header__meta">
|
<div className="page-header__meta">
|
||||||
@@ -468,14 +616,14 @@ function PackagePage() {
|
|||||||
</div>
|
</div>
|
||||||
{pkg && (pkg.tag_count !== undefined || pkg.artifact_count !== undefined) && (
|
{pkg && (pkg.tag_count !== undefined || pkg.artifact_count !== undefined) && (
|
||||||
<div className="package-header-stats">
|
<div className="package-header-stats">
|
||||||
{pkg.tag_count !== undefined && (
|
{!isSystemProject && pkg.tag_count !== undefined && (
|
||||||
<span className="stat-item">
|
<span className="stat-item">
|
||||||
<strong>{pkg.tag_count}</strong> tags
|
<strong>{pkg.tag_count}</strong> tags
|
||||||
</span>
|
</span>
|
||||||
)}
|
)}
|
||||||
{pkg.artifact_count !== undefined && (
|
{pkg.artifact_count !== undefined && (
|
||||||
<span className="stat-item">
|
<span className="stat-item">
|
||||||
<strong>{pkg.artifact_count}</strong> artifacts
|
<strong>{pkg.artifact_count}</strong> {isSystemProject ? 'versions' : 'artifacts'}
|
||||||
</span>
|
</span>
|
||||||
)}
|
)}
|
||||||
{pkg.total_size !== undefined && pkg.total_size > 0 && (
|
{pkg.total_size !== undefined && pkg.total_size > 0 && (
|
||||||
@@ -483,7 +631,7 @@ function PackagePage() {
|
|||||||
<strong>{formatBytes(pkg.total_size)}</strong> total
|
<strong>{formatBytes(pkg.total_size)}</strong> total
|
||||||
</span>
|
</span>
|
||||||
)}
|
)}
|
||||||
{pkg.latest_tag && (
|
{!isSystemProject && pkg.latest_tag && (
|
||||||
<span className="stat-item">
|
<span className="stat-item">
|
||||||
Latest: <strong className="accent">{pkg.latest_tag}</strong>
|
Latest: <strong className="accent">{pkg.latest_tag}</strong>
|
||||||
</span>
|
</span>
|
||||||
@@ -496,44 +644,9 @@ function PackagePage() {
|
|||||||
{error && <div className="error-message">{error}</div>}
|
{error && <div className="error-message">{error}</div>}
|
||||||
{uploadSuccess && <div className="success-message">{uploadSuccess}</div>}
|
{uploadSuccess && <div className="success-message">{uploadSuccess}</div>}
|
||||||
|
|
||||||
{user && (
|
|
||||||
<div className="upload-section card">
|
|
||||||
<h3>Upload Artifact</h3>
|
|
||||||
{canWrite ? (
|
|
||||||
<div className="upload-form">
|
|
||||||
<div className="form-group">
|
|
||||||
<label htmlFor="upload-tag">Tag (optional)</label>
|
|
||||||
<input
|
|
||||||
id="upload-tag"
|
|
||||||
type="text"
|
|
||||||
value={uploadTag}
|
|
||||||
onChange={(e) => setUploadTag(e.target.value)}
|
|
||||||
placeholder="v1.0.0, latest, stable..."
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
<DragDropUpload
|
|
||||||
projectName={projectName!}
|
|
||||||
packageName={packageName!}
|
|
||||||
tag={uploadTag || undefined}
|
|
||||||
onUploadComplete={handleUploadComplete}
|
|
||||||
onUploadError={handleUploadError}
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
) : (
|
|
||||||
<DragDropUpload
|
|
||||||
projectName={projectName!}
|
|
||||||
packageName={packageName!}
|
|
||||||
disabled={true}
|
|
||||||
disabledReason="You have read-only access to this project and cannot upload artifacts."
|
|
||||||
onUploadComplete={handleUploadComplete}
|
|
||||||
onUploadError={handleUploadError}
|
|
||||||
/>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
)}
|
|
||||||
|
|
||||||
<div className="section-header">
|
<div className="section-header">
|
||||||
<h2>Tags / Versions</h2>
|
<h2>{isSystemProject ? 'Versions' : 'Tags / Versions'}</h2>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div className="list-controls">
|
<div className="list-controls">
|
||||||
@@ -577,110 +690,6 @@ function PackagePage() {
|
|||||||
/>
|
/>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
{/* Dependencies Section */}
|
|
||||||
{tags.length > 0 && (
|
|
||||||
<div className="dependencies-section card">
|
|
||||||
<div className="dependencies-header">
|
|
||||||
<h3>Dependencies</h3>
|
|
||||||
<div className="dependencies-controls">
|
|
||||||
{selectedTag && (
|
|
||||||
<>
|
|
||||||
<button
|
|
||||||
className="btn btn-secondary btn-small"
|
|
||||||
onClick={fetchEnsureFile}
|
|
||||||
disabled={ensureFileLoading}
|
|
||||||
title="View orchard.ensure file"
|
|
||||||
>
|
|
||||||
<svg width="14" height="14" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2" style={{ marginRight: '6px' }}>
|
|
||||||
<path d="M14 2H6a2 2 0 0 0-2 2v16a2 2 0 0 0 2 2h12a2 2 0 0 0 2-2V8z"></path>
|
|
||||||
<polyline points="14 2 14 8 20 8"></polyline>
|
|
||||||
<line x1="16" y1="13" x2="8" y2="13"></line>
|
|
||||||
<line x1="16" y1="17" x2="8" y2="17"></line>
|
|
||||||
<polyline points="10 9 9 9 8 9"></polyline>
|
|
||||||
</svg>
|
|
||||||
{ensureFileLoading ? 'Loading...' : 'View Ensure File'}
|
|
||||||
</button>
|
|
||||||
<button
|
|
||||||
className="btn btn-secondary btn-small"
|
|
||||||
onClick={() => setShowGraph(true)}
|
|
||||||
title="View full dependency tree"
|
|
||||||
>
|
|
||||||
<svg width="14" height="14" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2" style={{ marginRight: '6px' }}>
|
|
||||||
<circle cx="12" cy="12" r="3"></circle>
|
|
||||||
<circle cx="4" cy="4" r="2"></circle>
|
|
||||||
<circle cx="20" cy="4" r="2"></circle>
|
|
||||||
<circle cx="4" cy="20" r="2"></circle>
|
|
||||||
<circle cx="20" cy="20" r="2"></circle>
|
|
||||||
<line x1="9.5" y1="9.5" x2="5.5" y2="5.5"></line>
|
|
||||||
<line x1="14.5" y1="9.5" x2="18.5" y2="5.5"></line>
|
|
||||||
<line x1="9.5" y1="14.5" x2="5.5" y2="18.5"></line>
|
|
||||||
<line x1="14.5" y1="14.5" x2="18.5" y2="18.5"></line>
|
|
||||||
</svg>
|
|
||||||
View Graph
|
|
||||||
</button>
|
|
||||||
</>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
<div className="dependencies-tag-select">
|
|
||||||
{selectedTag && (
|
|
||||||
<select
|
|
||||||
className="tag-selector"
|
|
||||||
value={selectedTag.id}
|
|
||||||
onChange={(e) => {
|
|
||||||
const tag = tags.find(t => t.id === e.target.value);
|
|
||||||
if (tag) setSelectedTag(tag);
|
|
||||||
}}
|
|
||||||
>
|
|
||||||
{tags.map(t => (
|
|
||||||
<option key={t.id} value={t.id}>
|
|
||||||
{t.name}{t.version ? ` (${t.version})` : ''}
|
|
||||||
</option>
|
|
||||||
))}
|
|
||||||
</select>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
|
|
||||||
{depsLoading ? (
|
|
||||||
<div className="deps-loading">Loading dependencies...</div>
|
|
||||||
) : depsError ? (
|
|
||||||
<div className="deps-error">{depsError}</div>
|
|
||||||
) : dependencies.length === 0 ? (
|
|
||||||
<div className="deps-empty">
|
|
||||||
{selectedTag ? (
|
|
||||||
<span><strong>{selectedTag.name}</strong> has no dependencies</span>
|
|
||||||
) : (
|
|
||||||
<span>No dependencies</span>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
) : (
|
|
||||||
<div className="deps-list">
|
|
||||||
<div className="deps-summary">
|
|
||||||
<strong>{selectedTag?.name}</strong> has {dependencies.length} {dependencies.length === 1 ? 'dependency' : 'dependencies'}:
|
|
||||||
</div>
|
|
||||||
<ul className="deps-items">
|
|
||||||
{dependencies.map((dep) => (
|
|
||||||
<li key={dep.id} className="dep-item">
|
|
||||||
<Link
|
|
||||||
to={`/project/${dep.project}/${dep.package}`}
|
|
||||||
className="dep-link"
|
|
||||||
>
|
|
||||||
{dep.project}/{dep.package}
|
|
||||||
</Link>
|
|
||||||
<span className="dep-constraint">
|
|
||||||
@ {dep.version || dep.tag}
|
|
||||||
</span>
|
|
||||||
<span className="dep-status dep-status--ok" title="Package exists">
|
|
||||||
✓
|
|
||||||
</span>
|
|
||||||
</li>
|
|
||||||
))}
|
|
||||||
</ul>
|
|
||||||
</div>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{/* Used By (Reverse Dependencies) Section */}
|
{/* Used By (Reverse Dependencies) Section */}
|
||||||
<div className="used-by-section card">
|
<div className="used-by-section card">
|
||||||
<h3>Used By</h3>
|
<h3>Used By</h3>
|
||||||
@@ -737,78 +746,6 @@ function PackagePage() {
|
|||||||
)}
|
)}
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div className="download-by-id-section card">
|
|
||||||
<h3>Download by Artifact ID</h3>
|
|
||||||
<div className="download-by-id-form">
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={artifactIdInput}
|
|
||||||
onChange={(e) => setArtifactIdInput(e.target.value.toLowerCase().replace(/[^a-f0-9]/g, '').slice(0, 64))}
|
|
||||||
placeholder="Enter SHA256 artifact ID (64 hex characters)"
|
|
||||||
className="artifact-id-input"
|
|
||||||
/>
|
|
||||||
<a
|
|
||||||
href={artifactIdInput.length === 64 ? getDownloadUrl(projectName!, packageName!, `artifact:${artifactIdInput}`) : '#'}
|
|
||||||
className={`btn btn-primary ${artifactIdInput.length !== 64 ? 'btn-disabled' : ''}`}
|
|
||||||
download
|
|
||||||
onClick={(e) => {
|
|
||||||
if (artifactIdInput.length !== 64) {
|
|
||||||
e.preventDefault();
|
|
||||||
}
|
|
||||||
}}
|
|
||||||
>
|
|
||||||
Download
|
|
||||||
</a>
|
|
||||||
</div>
|
|
||||||
{artifactIdInput.length > 0 && artifactIdInput.length !== 64 && (
|
|
||||||
<p className="validation-hint">Artifact ID must be exactly 64 hex characters ({artifactIdInput.length}/64)</p>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
|
|
||||||
{user && canWrite && (
|
|
||||||
<div className="create-tag-section card">
|
|
||||||
<h3>Create / Update Tag</h3>
|
|
||||||
<p className="section-description">Point a tag at any existing artifact by its ID</p>
|
|
||||||
<form onSubmit={handleCreateTag} className="create-tag-form">
|
|
||||||
<div className="form-row">
|
|
||||||
<div className="form-group">
|
|
||||||
<label htmlFor="create-tag-name">Tag Name</label>
|
|
||||||
<input
|
|
||||||
id="create-tag-name"
|
|
||||||
type="text"
|
|
||||||
value={createTagName}
|
|
||||||
onChange={(e) => setCreateTagName(e.target.value)}
|
|
||||||
placeholder="latest, stable, v1.0.0..."
|
|
||||||
disabled={createTagLoading}
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
<div className="form-group form-group--wide">
|
|
||||||
<label htmlFor="create-tag-artifact">Artifact ID</label>
|
|
||||||
<input
|
|
||||||
id="create-tag-artifact"
|
|
||||||
type="text"
|
|
||||||
value={createTagArtifactId}
|
|
||||||
onChange={(e) => setCreateTagArtifactId(e.target.value.toLowerCase().replace(/[^a-f0-9]/g, '').slice(0, 64))}
|
|
||||||
placeholder="SHA256 hash (64 hex characters)"
|
|
||||||
className="artifact-id-input"
|
|
||||||
disabled={createTagLoading}
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
<button
|
|
||||||
type="submit"
|
|
||||||
className="btn btn-primary"
|
|
||||||
disabled={createTagLoading || !createTagName.trim() || createTagArtifactId.length !== 64}
|
|
||||||
>
|
|
||||||
{createTagLoading ? 'Creating...' : 'Create Tag'}
|
|
||||||
</button>
|
|
||||||
</div>
|
|
||||||
{createTagArtifactId.length > 0 && createTagArtifactId.length !== 64 && (
|
|
||||||
<p className="validation-hint">Artifact ID must be exactly 64 hex characters ({createTagArtifactId.length}/64)</p>
|
|
||||||
)}
|
|
||||||
</form>
|
|
||||||
</div>
|
|
||||||
)}
|
|
||||||
|
|
||||||
<div className="usage-section card">
|
<div className="usage-section card">
|
||||||
<h3>Usage</h3>
|
<h3>Usage</h3>
|
||||||
<p>Download artifacts using:</p>
|
<p>Download artifacts using:</p>
|
||||||
@@ -831,6 +768,118 @@ function PackagePage() {
|
|||||||
/>
|
/>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
|
{/* Upload Modal */}
|
||||||
|
{showUploadModal && (
|
||||||
|
<div className="modal-overlay" onClick={() => setShowUploadModal(false)}>
|
||||||
|
<div className="upload-modal" onClick={(e) => e.stopPropagation()}>
|
||||||
|
<div className="modal-header">
|
||||||
|
<h3>Upload Artifact</h3>
|
||||||
|
<button
|
||||||
|
className="modal-close"
|
||||||
|
onClick={() => setShowUploadModal(false)}
|
||||||
|
title="Close"
|
||||||
|
>
|
||||||
|
<svg width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2">
|
||||||
|
<line x1="18" y1="6" x2="6" y2="18"></line>
|
||||||
|
<line x1="6" y1="6" x2="18" y2="18"></line>
|
||||||
|
</svg>
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
<div className="modal-body">
|
||||||
|
<div className="form-group">
|
||||||
|
<label htmlFor="upload-tag">Tag (optional)</label>
|
||||||
|
<input
|
||||||
|
id="upload-tag"
|
||||||
|
type="text"
|
||||||
|
value={uploadTag}
|
||||||
|
onChange={(e) => setUploadTag(e.target.value)}
|
||||||
|
placeholder="v1.0.0, latest, stable..."
|
||||||
|
/>
|
||||||
|
</div>
|
||||||
|
<DragDropUpload
|
||||||
|
projectName={projectName!}
|
||||||
|
packageName={packageName!}
|
||||||
|
tag={uploadTag || undefined}
|
||||||
|
onUploadComplete={(result) => {
|
||||||
|
handleUploadComplete(result);
|
||||||
|
setShowUploadModal(false);
|
||||||
|
setUploadTag('');
|
||||||
|
}}
|
||||||
|
onUploadError={handleUploadError}
|
||||||
|
/>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{/* Create/Update Tag Modal */}
|
||||||
|
{showCreateTagModal && (
|
||||||
|
<div className="modal-overlay" onClick={() => setShowCreateTagModal(false)}>
|
||||||
|
<div className="create-tag-modal" onClick={(e) => e.stopPropagation()}>
|
||||||
|
<div className="modal-header">
|
||||||
|
<h3>Create / Update Tag</h3>
|
||||||
|
<button
|
||||||
|
className="modal-close"
|
||||||
|
onClick={() => { setShowCreateTagModal(false); setCreateTagName(''); setCreateTagArtifactId(''); }}
|
||||||
|
title="Close"
|
||||||
|
>
|
||||||
|
<svg width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2">
|
||||||
|
<line x1="18" y1="6" x2="6" y2="18"></line>
|
||||||
|
<line x1="6" y1="6" x2="18" y2="18"></line>
|
||||||
|
</svg>
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
<div className="modal-body">
|
||||||
|
<p className="modal-description">Point a tag at an artifact by its ID</p>
|
||||||
|
<form onSubmit={(e) => { handleCreateTag(e); setShowCreateTagModal(false); }}>
|
||||||
|
<div className="form-group">
|
||||||
|
<label htmlFor="modal-tag-name">Tag Name</label>
|
||||||
|
<input
|
||||||
|
id="modal-tag-name"
|
||||||
|
type="text"
|
||||||
|
value={createTagName}
|
||||||
|
onChange={(e) => setCreateTagName(e.target.value)}
|
||||||
|
placeholder="latest, stable, v1.0.0..."
|
||||||
|
disabled={createTagLoading}
|
||||||
|
/>
|
||||||
|
</div>
|
||||||
|
<div className="form-group">
|
||||||
|
<label htmlFor="modal-artifact-id">Artifact ID</label>
|
||||||
|
<input
|
||||||
|
id="modal-artifact-id"
|
||||||
|
type="text"
|
||||||
|
value={createTagArtifactId}
|
||||||
|
onChange={(e) => setCreateTagArtifactId(e.target.value.toLowerCase().replace(/[^a-f0-9]/g, '').slice(0, 64))}
|
||||||
|
placeholder="SHA256 hash (64 hex characters)"
|
||||||
|
className="artifact-id-input"
|
||||||
|
disabled={createTagLoading}
|
||||||
|
/>
|
||||||
|
{createTagArtifactId.length > 0 && createTagArtifactId.length !== 64 && (
|
||||||
|
<p className="validation-hint">{createTagArtifactId.length}/64 characters</p>
|
||||||
|
)}
|
||||||
|
</div>
|
||||||
|
<div className="modal-actions">
|
||||||
|
<button
|
||||||
|
type="button"
|
||||||
|
className="btn btn-secondary"
|
||||||
|
onClick={() => { setShowCreateTagModal(false); setCreateTagName(''); setCreateTagArtifactId(''); }}
|
||||||
|
>
|
||||||
|
Cancel
|
||||||
|
</button>
|
||||||
|
<button
|
||||||
|
type="submit"
|
||||||
|
className="btn btn-primary"
|
||||||
|
disabled={createTagLoading || !createTagName.trim() || createTagArtifactId.length !== 64}
|
||||||
|
>
|
||||||
|
{createTagLoading ? 'Creating...' : 'Create Tag'}
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
</form>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
)}
|
||||||
|
|
||||||
{/* Ensure File Modal */}
|
{/* Ensure File Modal */}
|
||||||
{showEnsureFile && (
|
{showEnsureFile && (
|
||||||
<div className="modal-overlay" onClick={() => setShowEnsureFile(false)}>
|
<div className="modal-overlay" onClick={() => setShowEnsureFile(false)}>
|
||||||
@@ -872,6 +921,107 @@ function PackagePage() {
|
|||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
|
{/* Dependencies Modal */}
|
||||||
|
{showDepsModal && selectedTag && (
|
||||||
|
<div className="modal-overlay" onClick={() => setShowDepsModal(false)}>
|
||||||
|
<div className="deps-modal" onClick={(e) => e.stopPropagation()}>
|
||||||
|
<div className="modal-header">
|
||||||
|
<h3>Dependencies for {selectedTag.version || selectedTag.name}</h3>
|
||||||
|
<button
|
||||||
|
className="modal-close"
|
||||||
|
onClick={() => setShowDepsModal(false)}
|
||||||
|
title="Close"
|
||||||
|
>
|
||||||
|
<svg width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2">
|
||||||
|
<line x1="18" y1="6" x2="6" y2="18"></line>
|
||||||
|
<line x1="6" y1="6" x2="18" y2="18"></line>
|
||||||
|
</svg>
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
<div className="modal-body">
|
||||||
|
<div className="deps-modal-controls">
|
||||||
|
<button
|
||||||
|
className="btn btn-secondary btn-small"
|
||||||
|
onClick={fetchEnsureFile}
|
||||||
|
disabled={ensureFileLoading}
|
||||||
|
>
|
||||||
|
View Ensure File
|
||||||
|
</button>
|
||||||
|
<button
|
||||||
|
className="btn btn-secondary btn-small"
|
||||||
|
onClick={() => { setShowDepsModal(false); setShowGraph(true); }}
|
||||||
|
>
|
||||||
|
View Graph
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
{depsLoading ? (
|
||||||
|
<div className="deps-loading">Loading dependencies...</div>
|
||||||
|
) : depsError ? (
|
||||||
|
<div className="deps-error">{depsError}</div>
|
||||||
|
) : dependencies.length === 0 ? (
|
||||||
|
<div className="deps-empty">No dependencies</div>
|
||||||
|
) : (
|
||||||
|
<div className="deps-list">
|
||||||
|
<div className="deps-summary">
|
||||||
|
{dependencies.length} {dependencies.length === 1 ? 'dependency' : 'dependencies'}:
|
||||||
|
</div>
|
||||||
|
<ul className="deps-items">
|
||||||
|
{dependencies.map((dep) => (
|
||||||
|
<li key={dep.id} className="dep-item">
|
||||||
|
<Link
|
||||||
|
to={`/project/${dep.project}/${dep.package}`}
|
||||||
|
className="dep-link"
|
||||||
|
onClick={() => setShowDepsModal(false)}
|
||||||
|
>
|
||||||
|
{dep.project}/{dep.package}
|
||||||
|
</Link>
|
||||||
|
<span className="dep-constraint">
|
||||||
|
@ {dep.version || dep.tag}
|
||||||
|
</span>
|
||||||
|
<span className="dep-status dep-status--ok" title="Package exists">
|
||||||
|
✓
|
||||||
|
</span>
|
||||||
|
</li>
|
||||||
|
))}
|
||||||
|
</ul>
|
||||||
|
</div>
|
||||||
|
)}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{/* Artifact ID Modal */}
|
||||||
|
{showArtifactIdModal && viewArtifactId && (
|
||||||
|
<div className="modal-overlay" onClick={() => setShowArtifactIdModal(false)}>
|
||||||
|
<div className="artifact-id-modal" onClick={(e) => e.stopPropagation()}>
|
||||||
|
<div className="modal-header">
|
||||||
|
<h3>Artifact ID</h3>
|
||||||
|
<button
|
||||||
|
className="modal-close"
|
||||||
|
onClick={() => setShowArtifactIdModal(false)}
|
||||||
|
title="Close"
|
||||||
|
>
|
||||||
|
<svg width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" strokeWidth="2">
|
||||||
|
<line x1="18" y1="6" x2="6" y2="18"></line>
|
||||||
|
<line x1="6" y1="6" x2="18" y2="18"></line>
|
||||||
|
</svg>
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
<div className="modal-body">
|
||||||
|
<p className="modal-description">SHA256 hash identifying this artifact:</p>
|
||||||
|
<div className="artifact-id-display">
|
||||||
|
<code>{viewArtifactId}</code>
|
||||||
|
<CopyButton text={viewArtifactId} />
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{/* Action Menu Dropdown */}
|
||||||
|
{renderActionMenu()}
|
||||||
</div>
|
</div>
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -214,7 +214,7 @@ function ProjectPage() {
|
|||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
<div className="page-header__actions">
|
<div className="page-header__actions">
|
||||||
{canAdmin && !project.team_id && (
|
{canAdmin && !project.team_id && !project.is_system && (
|
||||||
<button
|
<button
|
||||||
className="btn btn-secondary"
|
className="btn btn-secondary"
|
||||||
onClick={() => navigate(`/project/${projectName}/settings`)}
|
onClick={() => navigate(`/project/${projectName}/settings`)}
|
||||||
@@ -227,11 +227,11 @@ function ProjectPage() {
|
|||||||
Settings
|
Settings
|
||||||
</button>
|
</button>
|
||||||
)}
|
)}
|
||||||
{canWrite ? (
|
{canWrite && !project.is_system ? (
|
||||||
<button className="btn btn-primary" onClick={() => setShowForm(!showForm)}>
|
<button className="btn btn-primary" onClick={() => setShowForm(!showForm)}>
|
||||||
{showForm ? 'Cancel' : '+ New Package'}
|
{showForm ? 'Cancel' : '+ New Package'}
|
||||||
</button>
|
</button>
|
||||||
) : user ? (
|
) : user && !project.is_system ? (
|
||||||
<span className="text-muted" title="You have read-only access to this project">
|
<span className="text-muted" title="You have read-only access to this project">
|
||||||
Read-only access
|
Read-only access
|
||||||
</span>
|
</span>
|
||||||
@@ -294,18 +294,20 @@ function ProjectPage() {
|
|||||||
placeholder="Filter packages..."
|
placeholder="Filter packages..."
|
||||||
className="list-controls__search"
|
className="list-controls__search"
|
||||||
/>
|
/>
|
||||||
<select
|
{!project?.is_system && (
|
||||||
className="list-controls__select"
|
<select
|
||||||
value={format}
|
className="list-controls__select"
|
||||||
onChange={(e) => handleFormatChange(e.target.value)}
|
value={format}
|
||||||
>
|
onChange={(e) => handleFormatChange(e.target.value)}
|
||||||
<option value="">All formats</option>
|
>
|
||||||
{FORMAT_OPTIONS.map((f) => (
|
<option value="">All formats</option>
|
||||||
<option key={f} value={f}>
|
{FORMAT_OPTIONS.map((f) => (
|
||||||
{f}
|
<option key={f} value={f}>
|
||||||
</option>
|
{f}
|
||||||
))}
|
</option>
|
||||||
</select>
|
))}
|
||||||
|
</select>
|
||||||
|
)}
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
{hasActiveFilters && (
|
{hasActiveFilters && (
|
||||||
@@ -341,19 +343,19 @@ function ProjectPage() {
|
|||||||
className: 'cell-description',
|
className: 'cell-description',
|
||||||
render: (pkg) => pkg.description || '—',
|
render: (pkg) => pkg.description || '—',
|
||||||
},
|
},
|
||||||
{
|
...(!project?.is_system ? [{
|
||||||
key: 'format',
|
key: 'format',
|
||||||
header: 'Format',
|
header: 'Format',
|
||||||
render: (pkg) => <Badge variant="default">{pkg.format}</Badge>,
|
render: (pkg: Package) => <Badge variant="default">{pkg.format}</Badge>,
|
||||||
},
|
}] : []),
|
||||||
{
|
...(!project?.is_system ? [{
|
||||||
key: 'tag_count',
|
key: 'tag_count',
|
||||||
header: 'Tags',
|
header: 'Tags',
|
||||||
render: (pkg) => pkg.tag_count ?? '—',
|
render: (pkg: Package) => pkg.tag_count ?? '—',
|
||||||
},
|
}] : []),
|
||||||
{
|
{
|
||||||
key: 'artifact_count',
|
key: 'artifact_count',
|
||||||
header: 'Artifacts',
|
header: project?.is_system ? 'Versions' : 'Artifacts',
|
||||||
render: (pkg) => pkg.artifact_count ?? '—',
|
render: (pkg) => pkg.artifact_count ?? '—',
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@@ -362,12 +364,12 @@ function ProjectPage() {
|
|||||||
render: (pkg) =>
|
render: (pkg) =>
|
||||||
pkg.total_size !== undefined && pkg.total_size > 0 ? formatBytes(pkg.total_size) : '—',
|
pkg.total_size !== undefined && pkg.total_size > 0 ? formatBytes(pkg.total_size) : '—',
|
||||||
},
|
},
|
||||||
{
|
...(!project?.is_system ? [{
|
||||||
key: 'latest_tag',
|
key: 'latest_tag',
|
||||||
header: 'Latest',
|
header: 'Latest',
|
||||||
render: (pkg) =>
|
render: (pkg: Package) =>
|
||||||
pkg.latest_tag ? <strong style={{ color: 'var(--accent-primary)' }}>{pkg.latest_tag}</strong> : '—',
|
pkg.latest_tag ? <strong style={{ color: 'var(--accent-primary)' }}>{pkg.latest_tag}</strong> : '—',
|
||||||
},
|
}] : []),
|
||||||
{
|
{
|
||||||
key: 'created_at',
|
key: 'created_at',
|
||||||
header: 'Created',
|
header: 'Created',
|
||||||
|
|||||||
Reference in New Issue
Block a user