Compare commits
6 Commits
feature/cp
...
dca05a8e9d
| Author | SHA1 | Date | |
|---|---|---|---|
| dca05a8e9d | |||
| e858fea056 | |||
| 90a4264691 | |||
| a94ba0d8a0 | |||
| 7253c64c78 | |||
| b074b8fe47 |
@@ -1,10 +0,0 @@
|
|||||||
.git
|
|
||||||
bin/
|
|
||||||
.editorconfig
|
|
||||||
.vscode/
|
|
||||||
.idea/
|
|
||||||
*.test
|
|
||||||
LICENSE
|
|
||||||
CONVENTIONS.md
|
|
||||||
REPO_POLICIES.md
|
|
||||||
README.md
|
|
||||||
@@ -1,15 +0,0 @@
|
|||||||
root = true
|
|
||||||
|
|
||||||
[*]
|
|
||||||
charset = utf-8
|
|
||||||
end_of_line = lf
|
|
||||||
insert_final_newline = true
|
|
||||||
trim_trailing_whitespace = true
|
|
||||||
indent_style = space
|
|
||||||
indent_size = 2
|
|
||||||
|
|
||||||
[*.go]
|
|
||||||
indent_style = tab
|
|
||||||
|
|
||||||
[Makefile]
|
|
||||||
indent_style = tab
|
|
||||||
@@ -10,7 +10,17 @@ jobs:
|
|||||||
check:
|
check:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@34e114876b0b11c390a56381ad16ebd13914f8d5 # v4, 2024-10-13
|
- uses: actions/checkout@34e114876b0b11c390a56381ad16ebd13914f8d5 # v4
|
||||||
|
|
||||||
- name: Build (runs make check inside Dockerfile)
|
- uses: actions/setup-go@40f1582b2485089dde7abd97c1529aa768e1baff # v5
|
||||||
run: docker build .
|
with:
|
||||||
|
go-version-file: go.mod
|
||||||
|
|
||||||
|
- name: Install golangci-lint
|
||||||
|
run: go install github.com/golangci/golangci-lint/v2/cmd/golangci-lint@5d1e709b7be35cb2025444e19de266b056b7b7ee # v2.10.1
|
||||||
|
|
||||||
|
- name: Install goimports
|
||||||
|
run: go install golang.org/x/tools/cmd/goimports@009367f5c17a8d4c45a961a3a509277190a9a6f0 # v0.42.0
|
||||||
|
|
||||||
|
- name: Run make check
|
||||||
|
run: make check
|
||||||
|
|||||||
31
.gitignore
vendored
31
.gitignore
vendored
@@ -1,31 +0,0 @@
|
|||||||
# OS
|
|
||||||
.DS_Store
|
|
||||||
Thumbs.db
|
|
||||||
|
|
||||||
# Editors
|
|
||||||
*.swp
|
|
||||||
*.swo
|
|
||||||
*~
|
|
||||||
*.bak
|
|
||||||
.idea/
|
|
||||||
.vscode/
|
|
||||||
*.sublime-*
|
|
||||||
|
|
||||||
# Node
|
|
||||||
node_modules/
|
|
||||||
|
|
||||||
# Environment / secrets
|
|
||||||
.env
|
|
||||||
.env.*
|
|
||||||
*.pem
|
|
||||||
*.key
|
|
||||||
|
|
||||||
# Go
|
|
||||||
bin/
|
|
||||||
*.exe
|
|
||||||
*.exe~
|
|
||||||
*.dll
|
|
||||||
*.so
|
|
||||||
*.dylib
|
|
||||||
*.test
|
|
||||||
*.out
|
|
||||||
33
Dockerfile
33
Dockerfile
@@ -1,37 +1,26 @@
|
|||||||
# Lint stage — fast feedback on formatting and lint issues
|
# Build stage
|
||||||
# golangci/golangci-lint:v2.10.1
|
FROM golang:1.25-alpine AS builder
|
||||||
FROM golangci/golangci-lint@sha256:ea84d14c2fef724411be7dc45e09e6ef721d748315252b02df19a7e3113ee763 AS lint
|
|
||||||
|
|
||||||
WORKDIR /src
|
|
||||||
COPY go.mod go.sum ./
|
|
||||||
RUN go mod download
|
|
||||||
|
|
||||||
COPY . .
|
|
||||||
|
|
||||||
RUN make fmt-check
|
|
||||||
RUN make lint
|
|
||||||
|
|
||||||
# Build stage — tests and compilation
|
|
||||||
# golang:1.25-alpine
|
|
||||||
FROM golang@sha256:f6751d823c26342f9506c03797d2527668d095b0a15f1862cddb4d927a7a4ced AS builder
|
|
||||||
|
|
||||||
# Force BuildKit to run the lint stage by creating a stage dependency
|
|
||||||
COPY --from=lint /src/go.sum /dev/null
|
|
||||||
|
|
||||||
RUN apk add --no-cache git make gcc musl-dev
|
RUN apk add --no-cache git make gcc musl-dev
|
||||||
|
|
||||||
|
# Install golangci-lint v2
|
||||||
|
RUN go install github.com/golangci/golangci-lint/v2/cmd/golangci-lint@latest
|
||||||
|
RUN go install golang.org/x/tools/cmd/goimports@latest
|
||||||
|
|
||||||
WORKDIR /src
|
WORKDIR /src
|
||||||
COPY go.mod go.sum ./
|
COPY go.mod go.sum ./
|
||||||
RUN go mod download
|
RUN go mod download
|
||||||
|
|
||||||
COPY . .
|
COPY . .
|
||||||
|
|
||||||
RUN make test
|
# Run all checks - build fails if any check fails
|
||||||
|
RUN make check
|
||||||
|
|
||||||
|
# Build the binary
|
||||||
RUN make build
|
RUN make build
|
||||||
|
|
||||||
# Runtime stage
|
# Runtime stage
|
||||||
# alpine:3.19
|
FROM alpine:3.19
|
||||||
FROM alpine@sha256:6baf43584bcb78f2e5847d1de515f23499913ac9f12bdf834811a3145eb11ca1
|
|
||||||
|
|
||||||
RUN apk add --no-cache ca-certificates tzdata git openssh-client docker-cli
|
RUN apk add --no-cache ca-certificates tzdata git openssh-client docker-cli
|
||||||
|
|
||||||
|
|||||||
27
Makefile
27
Makefile
@@ -1,4 +1,4 @@
|
|||||||
.PHONY: all build lint fmt fmt-check test check clean docker hooks
|
.PHONY: all build lint fmt test check clean
|
||||||
|
|
||||||
BINARY := upaasd
|
BINARY := upaasd
|
||||||
VERSION := $(shell git describe --tags --always --dirty 2>/dev/null || echo "dev")
|
VERSION := $(shell git describe --tags --always --dirty 2>/dev/null || echo "dev")
|
||||||
@@ -18,26 +18,21 @@ fmt:
|
|||||||
goimports -w .
|
goimports -w .
|
||||||
npx prettier --write --tab-width 4 static/js/*.js
|
npx prettier --write --tab-width 4 static/js/*.js
|
||||||
|
|
||||||
fmt-check:
|
|
||||||
@test -z "$$(gofmt -l .)" || (echo "Files not formatted:" && gofmt -l . && exit 1)
|
|
||||||
|
|
||||||
test:
|
test:
|
||||||
go test -v -race -cover -timeout 30s ./...
|
go test -v -race -cover ./...
|
||||||
|
|
||||||
# Check runs all validation without making changes
|
# Check runs all validation without making changes
|
||||||
# Used by CI and Docker build - fails if anything is wrong
|
# Used by CI and Docker build - fails if anything is wrong
|
||||||
check: fmt-check lint test
|
check:
|
||||||
|
@echo "==> Checking formatting..."
|
||||||
|
@test -z "$$(gofmt -l .)" || (echo "Files not formatted:" && gofmt -l . && exit 1)
|
||||||
|
@echo "==> Running linter..."
|
||||||
|
golangci-lint run --config .golangci.yml ./...
|
||||||
|
@echo "==> Running tests..."
|
||||||
|
go test -v -race ./...
|
||||||
|
@echo "==> Building..."
|
||||||
|
go build -ldflags "$(LDFLAGS)" -o /dev/null ./cmd/upaasd
|
||||||
@echo "==> All checks passed!"
|
@echo "==> All checks passed!"
|
||||||
|
|
||||||
docker:
|
|
||||||
docker build .
|
|
||||||
|
|
||||||
hooks:
|
|
||||||
@echo "Installing pre-commit hook..."
|
|
||||||
@mkdir -p .git/hooks
|
|
||||||
@printf '#!/bin/sh\nmake check\n' > .git/hooks/pre-commit
|
|
||||||
@chmod +x .git/hooks/pre-commit
|
|
||||||
@echo "Pre-commit hook installed."
|
|
||||||
|
|
||||||
clean:
|
clean:
|
||||||
rm -rf bin/
|
rm -rf bin/
|
||||||
|
|||||||
43
README.md
43
README.md
@@ -9,7 +9,6 @@ A simple self-hosted PaaS that auto-deploys Docker containers from Git repositor
|
|||||||
- Per-app UUID-based webhook URLs for Gitea integration
|
- Per-app UUID-based webhook URLs for Gitea integration
|
||||||
- Branch filtering - only deploy on configured branch changes
|
- Branch filtering - only deploy on configured branch changes
|
||||||
- Environment variables, labels, and volume mounts per app
|
- Environment variables, labels, and volume mounts per app
|
||||||
- CPU and memory resource limits per app
|
|
||||||
- Docker builds via socket access
|
- Docker builds via socket access
|
||||||
- Notifications via ntfy and Slack-compatible webhooks
|
- Notifications via ntfy and Slack-compatible webhooks
|
||||||
- Simple server-rendered UI with Tailwind CSS
|
- Simple server-rendered UI with Tailwind CSS
|
||||||
@@ -112,13 +111,10 @@ chi Router ──► Middleware Stack ──► Handler
|
|||||||
|
|
||||||
```bash
|
```bash
|
||||||
make fmt # Format code
|
make fmt # Format code
|
||||||
make fmt-check # Check formatting (read-only, fails if unformatted)
|
|
||||||
make lint # Run comprehensive linting
|
make lint # Run comprehensive linting
|
||||||
make test # Run tests with race detection (30s timeout)
|
make test # Run tests with race detection
|
||||||
make check # Verify everything passes (fmt-check, lint, test)
|
make check # Verify everything passes (lint, test, build, format)
|
||||||
make build # Build binary
|
make build # Build binary
|
||||||
make docker # Build Docker image
|
|
||||||
make hooks # Install pre-commit hook (runs make check)
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### Commit Requirements
|
### Commit Requirements
|
||||||
@@ -161,8 +157,8 @@ Environment variables:
|
|||||||
| Variable | Description | Default |
|
| Variable | Description | Default |
|
||||||
|----------|-------------|---------|
|
|----------|-------------|---------|
|
||||||
| `PORT` | HTTP listen port | 8080 |
|
| `PORT` | HTTP listen port | 8080 |
|
||||||
| `UPAAS_DATA_DIR` | Data directory for SQLite and keys | `./data` (local dev only — use absolute path for Docker) |
|
| `UPAAS_DATA_DIR` | Data directory for SQLite and keys | ./data |
|
||||||
| `UPAAS_HOST_DATA_DIR` | Host path for DATA_DIR (when running in container) | *(none — must be set to an absolute path)* |
|
| `UPAAS_HOST_DATA_DIR` | Host path for DATA_DIR (when running in container) | same as DATA_DIR |
|
||||||
| `UPAAS_DOCKER_HOST` | Docker socket path | unix:///var/run/docker.sock |
|
| `UPAAS_DOCKER_HOST` | Docker socket path | unix:///var/run/docker.sock |
|
||||||
| `DEBUG` | Enable debug logging | false |
|
| `DEBUG` | Enable debug logging | false |
|
||||||
| `SENTRY_DSN` | Sentry error reporting DSN | "" |
|
| `SENTRY_DSN` | Sentry error reporting DSN | "" |
|
||||||
@@ -180,35 +176,8 @@ docker run -d \
|
|||||||
upaas
|
upaas
|
||||||
```
|
```
|
||||||
|
|
||||||
### Docker Compose
|
**Important**: When running µPaaS inside a container, set `UPAAS_HOST_DATA_DIR` to the host path
|
||||||
|
that maps to `UPAAS_DATA_DIR`. This is required for Docker bind mounts during builds to work correctly.
|
||||||
```yaml
|
|
||||||
services:
|
|
||||||
upaas:
|
|
||||||
build: .
|
|
||||||
restart: unless-stopped
|
|
||||||
ports:
|
|
||||||
- "8080:8080"
|
|
||||||
volumes:
|
|
||||||
- /var/run/docker.sock:/var/run/docker.sock
|
|
||||||
- ${HOST_DATA_DIR}:/var/lib/upaas
|
|
||||||
environment:
|
|
||||||
- UPAAS_HOST_DATA_DIR=${HOST_DATA_DIR}
|
|
||||||
# Optional: uncomment to enable debug logging
|
|
||||||
# - DEBUG=true
|
|
||||||
# Optional: Sentry error reporting
|
|
||||||
# - SENTRY_DSN=https://...
|
|
||||||
# Optional: Prometheus metrics auth
|
|
||||||
# - METRICS_USERNAME=prometheus
|
|
||||||
# - METRICS_PASSWORD=secret
|
|
||||||
```
|
|
||||||
|
|
||||||
**Important**: You **must** set `HOST_DATA_DIR` to an **absolute path** on the host before running
|
|
||||||
`docker compose up`. This value is bind-mounted into the container and passed as `UPAAS_HOST_DATA_DIR`
|
|
||||||
so that Docker bind mounts during builds resolve correctly. Relative paths (e.g. `./data`) will break
|
|
||||||
container builds because the Docker daemon resolves paths relative to the host, not the container.
|
|
||||||
|
|
||||||
Example: `HOST_DATA_DIR=/srv/upaas/data docker compose up -d`
|
|
||||||
|
|
||||||
Session secrets are automatically generated on first startup and persisted to `$UPAAS_DATA_DIR/session.key`.
|
Session secrets are automatically generated on first startup and persisted to `$UPAAS_DATA_DIR/session.key`.
|
||||||
|
|
||||||
|
|||||||
188
REPO_POLICIES.md
188
REPO_POLICIES.md
@@ -1,188 +0,0 @@
|
|||||||
---
|
|
||||||
title: Repository Policies
|
|
||||||
last_modified: 2026-02-22
|
|
||||||
---
|
|
||||||
|
|
||||||
This document covers repository structure, tooling, and workflow standards. Code
|
|
||||||
style conventions are in separate documents:
|
|
||||||
|
|
||||||
- [Code Styleguide](https://git.eeqj.de/sneak/prompts/raw/branch/main/prompts/CODE_STYLEGUIDE.md)
|
|
||||||
(general, bash, Docker)
|
|
||||||
- [Go](https://git.eeqj.de/sneak/prompts/raw/branch/main/prompts/CODE_STYLEGUIDE_GO.md)
|
|
||||||
- [JavaScript](https://git.eeqj.de/sneak/prompts/raw/branch/main/prompts/CODE_STYLEGUIDE_JS.md)
|
|
||||||
- [Python](https://git.eeqj.de/sneak/prompts/raw/branch/main/prompts/CODE_STYLEGUIDE_PYTHON.md)
|
|
||||||
- [Go HTTP Server Conventions](https://git.eeqj.de/sneak/prompts/raw/branch/main/prompts/GO_HTTP_SERVER_CONVENTIONS.md)
|
|
||||||
|
|
||||||
---
|
|
||||||
|
|
||||||
- Cross-project documentation (such as this file) must include
|
|
||||||
`last_modified: YYYY-MM-DD` in the YAML front matter so it can be kept in sync
|
|
||||||
with the authoritative source as policies evolve.
|
|
||||||
|
|
||||||
- **ALL external references must be pinned by cryptographic hash.** This
|
|
||||||
includes Docker base images, Go modules, npm packages, GitHub Actions, and
|
|
||||||
anything else fetched from a remote source. Version tags (`@v4`, `@latest`,
|
|
||||||
`:3.21`, etc.) are server-mutable and therefore remote code execution
|
|
||||||
vulnerabilities. The ONLY acceptable way to reference an external dependency
|
|
||||||
is by its content hash (Docker `@sha256:...`, Go module hash in `go.sum`, npm
|
|
||||||
integrity hash in lockfile, GitHub Actions `@<commit-sha>`). No exceptions.
|
|
||||||
This also means never `curl | bash` to install tools like pyenv, nvm, rustup,
|
|
||||||
etc. Instead, download a specific release archive from GitHub, verify its hash
|
|
||||||
(hardcoded in the Dockerfile or script), and only then install. Unverified
|
|
||||||
install scripts are arbitrary remote code execution. This is the single most
|
|
||||||
important rule in this document. Double-check every external reference in
|
|
||||||
every file before committing. There are zero exceptions to this rule.
|
|
||||||
|
|
||||||
- Every repo with software must have a root `Makefile` with these targets:
|
|
||||||
`make test`, `make lint`, `make fmt` (writes), `make fmt-check` (read-only),
|
|
||||||
`make check` (prereqs: `test`, `lint`, `fmt-check`), `make docker`, and
|
|
||||||
`make hooks` (installs pre-commit hook). A model Makefile is at
|
|
||||||
`https://git.eeqj.de/sneak/prompts/raw/branch/main/Makefile`.
|
|
||||||
|
|
||||||
- Always use Makefile targets (`make fmt`, `make test`, `make lint`, etc.)
|
|
||||||
instead of invoking the underlying tools directly. The Makefile is the single
|
|
||||||
source of truth for how these operations are run.
|
|
||||||
|
|
||||||
- The Makefile is authoritative documentation for how the repo is used. Beyond
|
|
||||||
the required targets above, it should have targets for every common operation:
|
|
||||||
running a local development server (`make run`, `make dev`), re-initializing
|
|
||||||
or migrating the database (`make db-reset`, `make migrate`), building
|
|
||||||
artifacts (`make build`), generating code, seeding data, or anything else a
|
|
||||||
developer would do regularly. If someone checks out the repo and types
|
|
||||||
`make<tab>`, they should see every meaningful operation available. A new
|
|
||||||
contributor should be able to understand the entire development workflow by
|
|
||||||
reading the Makefile.
|
|
||||||
|
|
||||||
- Every repo should have a `Dockerfile`. All Dockerfiles must run `make check`
|
|
||||||
as a build step so the build fails if the branch is not green. For non-server
|
|
||||||
repos, the Dockerfile should bring up a development environment and run
|
|
||||||
`make check`. For server repos, `make check` should run as an early build
|
|
||||||
stage before the final image is assembled.
|
|
||||||
|
|
||||||
- Every repo should have a Gitea Actions workflow (`.gitea/workflows/`) that
|
|
||||||
runs `docker build .` on push. Since the Dockerfile already runs `make check`,
|
|
||||||
a successful build implies all checks pass.
|
|
||||||
|
|
||||||
- Use platform-standard formatters: `black` for Python, `prettier` for
|
|
||||||
JS/CSS/Markdown/HTML, `go fmt` for Go. Always use default configuration with
|
|
||||||
two exceptions: four-space indents (except Go), and `proseWrap: always` for
|
|
||||||
Markdown (hard-wrap at 80 columns). Documentation and writing repos (Markdown,
|
|
||||||
HTML, CSS) should also have `.prettierrc` and `.prettierignore`.
|
|
||||||
|
|
||||||
- Pre-commit hook: `make check` if local testing is possible, otherwise
|
|
||||||
`make lint && make fmt-check`. The Makefile should provide a `make hooks`
|
|
||||||
target to install the pre-commit hook.
|
|
||||||
|
|
||||||
- All repos with software must have tests that run via the platform-standard
|
|
||||||
test framework (`go test`, `pytest`, `jest`/`vitest`, etc.). If no meaningful
|
|
||||||
tests exist yet, add the most minimal test possible — e.g. importing the
|
|
||||||
module under test to verify it compiles/parses. There is no excuse for
|
|
||||||
`make test` to be a no-op.
|
|
||||||
|
|
||||||
- `make test` must complete in under 20 seconds. Add a 30-second timeout in the
|
|
||||||
Makefile.
|
|
||||||
|
|
||||||
- Docker builds must complete in under 5 minutes.
|
|
||||||
|
|
||||||
- `make check` must not modify any files in the repo. Tests may use temporary
|
|
||||||
directories.
|
|
||||||
|
|
||||||
- `main` must always pass `make check`, no exceptions.
|
|
||||||
|
|
||||||
- Never commit secrets. `.env` files, credentials, API keys, and private keys
|
|
||||||
must be in `.gitignore`. No exceptions.
|
|
||||||
|
|
||||||
- `.gitignore` should be comprehensive from the start: OS files (`.DS_Store`),
|
|
||||||
editor files (`.swp`, `*~`), language build artifacts, and `node_modules/`.
|
|
||||||
Fetch the standard `.gitignore` from
|
|
||||||
`https://git.eeqj.de/sneak/prompts/raw/branch/main/.gitignore` when setting up
|
|
||||||
a new repo.
|
|
||||||
|
|
||||||
- Never use `git add -A` or `git add .`. Always stage files explicitly by name.
|
|
||||||
|
|
||||||
- Never force-push to `main`.
|
|
||||||
|
|
||||||
- Make all changes on a feature branch. You can do whatever you want on a
|
|
||||||
feature branch.
|
|
||||||
|
|
||||||
- `.golangci.yml` is standardized and must _NEVER_ be modified by an agent, only
|
|
||||||
manually by the user. Fetch from
|
|
||||||
`https://git.eeqj.de/sneak/prompts/raw/branch/main/.golangci.yml`.
|
|
||||||
|
|
||||||
- When pinning images or packages by hash, add a comment above the reference
|
|
||||||
with the version and date (YYYY-MM-DD).
|
|
||||||
|
|
||||||
- Use `yarn`, not `npm`.
|
|
||||||
|
|
||||||
- Write all dates as YYYY-MM-DD (ISO 8601).
|
|
||||||
|
|
||||||
- Simple projects should be configured with environment variables.
|
|
||||||
|
|
||||||
- Dockerized web services listen on port 8080 by default, overridable with
|
|
||||||
`PORT`.
|
|
||||||
|
|
||||||
- `README.md` is the primary documentation. Required sections:
|
|
||||||
- **Description**: First line must include the project name, purpose,
|
|
||||||
category (web server, SPA, CLI tool, etc.), license, and author. Example:
|
|
||||||
"µPaaS is an MIT-licensed Go web application by @sneak that receives
|
|
||||||
git-frontend webhooks and deploys applications via Docker in realtime."
|
|
||||||
- **Getting Started**: Copy-pasteable install/usage code block.
|
|
||||||
- **Rationale**: Why does this exist?
|
|
||||||
- **Design**: How is the program structured?
|
|
||||||
- **TODO**: Update meticulously, even between commits. When planning, put
|
|
||||||
the todo list in the README so a new agent can pick up where the last one
|
|
||||||
left off.
|
|
||||||
- **License**: MIT, GPL, or WTFPL. Ask the user for new projects. Include a
|
|
||||||
`LICENSE` file in the repo root and a License section in the README.
|
|
||||||
- **Author**: [@sneak](https://sneak.berlin).
|
|
||||||
|
|
||||||
- First commit of a new repo should contain only `README.md`.
|
|
||||||
|
|
||||||
- Go module root: `sneak.berlin/go/<name>`. Always run `go mod tidy` before
|
|
||||||
committing.
|
|
||||||
|
|
||||||
- Use SemVer.
|
|
||||||
|
|
||||||
- Database migrations live in `internal/db/migrations/` and must be embedded in
|
|
||||||
the binary.
|
|
||||||
- `000_migration.sql` — contains ONLY the creation of the migrations tracking
|
|
||||||
table itself. Nothing else.
|
|
||||||
- `001_schema.sql` — the full application schema.
|
|
||||||
- **Pre-1.0.0:** never add additional migration files (002, 003, etc.). There
|
|
||||||
is no installed base to migrate. Edit `001_schema.sql` directly.
|
|
||||||
- **Post-1.0.0:** add new numbered migration files for each schema change.
|
|
||||||
Never edit existing migrations after release.
|
|
||||||
|
|
||||||
- All repos should have an `.editorconfig` enforcing the project's indentation
|
|
||||||
settings.
|
|
||||||
|
|
||||||
- Avoid putting files in the repo root unless necessary. Root should contain
|
|
||||||
only project-level config files (`README.md`, `Makefile`, `Dockerfile`,
|
|
||||||
`LICENSE`, `.gitignore`, `.editorconfig`, `REPO_POLICIES.md`, and
|
|
||||||
language-specific config). Everything else goes in a subdirectory. Canonical
|
|
||||||
subdirectory names:
|
|
||||||
- `bin/` — executable scripts and tools
|
|
||||||
- `cmd/` — Go command entrypoints
|
|
||||||
- `configs/` — configuration templates and examples
|
|
||||||
- `deploy/` — deployment manifests (k8s, compose, terraform)
|
|
||||||
- `docs/` — documentation and markdown (README.md stays in root)
|
|
||||||
- `internal/` — Go internal packages
|
|
||||||
- `internal/db/migrations/` — database migrations
|
|
||||||
- `pkg/` — Go library packages
|
|
||||||
- `share/` — systemd units, data files
|
|
||||||
- `static/` — static assets (images, fonts, etc.)
|
|
||||||
- `web/` — web frontend source
|
|
||||||
|
|
||||||
- When setting up a new repo, files from the `prompts` repo may be used as
|
|
||||||
templates. Fetch them from
|
|
||||||
`https://git.eeqj.de/sneak/prompts/raw/branch/main/<path>`.
|
|
||||||
|
|
||||||
- New repos must contain at minimum:
|
|
||||||
- `README.md`, `.git`, `.gitignore`, `.editorconfig`
|
|
||||||
- `LICENSE`, `REPO_POLICIES.md` (copy from the `prompts` repo)
|
|
||||||
- `Makefile`
|
|
||||||
- `Dockerfile`, `.dockerignore`
|
|
||||||
- `.gitea/workflows/check.yml`
|
|
||||||
- Go: `go.mod`, `go.sum`, `.golangci.yml`
|
|
||||||
- JS: `package.json`, `yarn.lock`, `.prettierrc`, `.prettierignore`
|
|
||||||
- Python: `pyproject.toml`
|
|
||||||
@@ -4,20 +4,20 @@ package main
|
|||||||
import (
|
import (
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/docker"
|
"git.eeqj.de/sneak/upaas/internal/docker"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/handlers"
|
"git.eeqj.de/sneak/upaas/internal/handlers"
|
||||||
"sneak.berlin/go/upaas/internal/healthcheck"
|
"git.eeqj.de/sneak/upaas/internal/healthcheck"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/middleware"
|
"git.eeqj.de/sneak/upaas/internal/middleware"
|
||||||
"sneak.berlin/go/upaas/internal/server"
|
"git.eeqj.de/sneak/upaas/internal/server"
|
||||||
"sneak.berlin/go/upaas/internal/service/app"
|
"git.eeqj.de/sneak/upaas/internal/service/app"
|
||||||
"sneak.berlin/go/upaas/internal/service/auth"
|
"git.eeqj.de/sneak/upaas/internal/service/auth"
|
||||||
"sneak.berlin/go/upaas/internal/service/deploy"
|
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
||||||
"sneak.berlin/go/upaas/internal/service/notify"
|
"git.eeqj.de/sneak/upaas/internal/service/notify"
|
||||||
"sneak.berlin/go/upaas/internal/service/webhook"
|
"git.eeqj.de/sneak/upaas/internal/service/webhook"
|
||||||
|
|
||||||
_ "github.com/joho/godotenv/autoload"
|
_ "github.com/joho/godotenv/autoload"
|
||||||
)
|
)
|
||||||
|
|||||||
18
docker-compose.yml
Normal file
18
docker-compose.yml
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
services:
|
||||||
|
upaas:
|
||||||
|
build: .
|
||||||
|
restart: unless-stopped
|
||||||
|
ports:
|
||||||
|
- "8080:8080"
|
||||||
|
volumes:
|
||||||
|
- /var/run/docker.sock:/var/run/docker.sock
|
||||||
|
- ${HOST_DATA_DIR:-./data}:/var/lib/upaas
|
||||||
|
environment:
|
||||||
|
- UPAAS_HOST_DATA_DIR=${HOST_DATA_DIR:-./data}
|
||||||
|
# Optional: uncomment to enable debug logging
|
||||||
|
# - DEBUG=true
|
||||||
|
# Optional: Sentry error reporting
|
||||||
|
# - SENTRY_DSN=https://...
|
||||||
|
# Optional: Prometheus metrics auth
|
||||||
|
# - METRICS_USERNAME=prometheus
|
||||||
|
# - METRICS_PASSWORD=secret
|
||||||
4
go.mod
4
go.mod
@@ -1,4 +1,4 @@
|
|||||||
module sneak.berlin/go/upaas
|
module git.eeqj.de/sneak/upaas
|
||||||
|
|
||||||
go 1.25
|
go 1.25
|
||||||
|
|
||||||
@@ -19,7 +19,6 @@ require (
|
|||||||
github.com/stretchr/testify v1.11.1
|
github.com/stretchr/testify v1.11.1
|
||||||
go.uber.org/fx v1.24.0
|
go.uber.org/fx v1.24.0
|
||||||
golang.org/x/crypto v0.46.0
|
golang.org/x/crypto v0.46.0
|
||||||
golang.org/x/time v0.12.0
|
|
||||||
)
|
)
|
||||||
|
|
||||||
require (
|
require (
|
||||||
@@ -75,6 +74,7 @@ require (
|
|||||||
go.yaml.in/yaml/v3 v3.0.4 // indirect
|
go.yaml.in/yaml/v3 v3.0.4 // indirect
|
||||||
golang.org/x/sys v0.39.0 // indirect
|
golang.org/x/sys v0.39.0 // indirect
|
||||||
golang.org/x/text v0.32.0 // indirect
|
golang.org/x/text v0.32.0 // indirect
|
||||||
|
golang.org/x/time v0.12.0 // indirect
|
||||||
google.golang.org/protobuf v1.36.10 // indirect
|
google.golang.org/protobuf v1.36.10 // indirect
|
||||||
gopkg.in/yaml.v3 v3.0.1 // indirect
|
gopkg.in/yaml.v3 v3.0.1 // indirect
|
||||||
gotest.tools/v3 v3.5.2 // indirect
|
gotest.tools/v3 v3.5.2 // indirect
|
||||||
|
|||||||
@@ -13,8 +13,8 @@ import (
|
|||||||
"github.com/spf13/viper"
|
"github.com/spf13/viper"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
// defaultPort is the default HTTP server port.
|
// defaultPort is the default HTTP server port.
|
||||||
|
|||||||
@@ -14,8 +14,8 @@ import (
|
|||||||
_ "github.com/mattn/go-sqlite3" // SQLite driver
|
_ "github.com/mattn/go-sqlite3" // SQLite driver
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
// dataDirPermissions is the file permission for the data directory.
|
// dataDirPermissions is the file permission for the data directory.
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import (
|
|||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestHashWebhookSecret(t *testing.T) {
|
func TestHashWebhookSecret(t *testing.T) {
|
||||||
|
|||||||
@@ -1,3 +0,0 @@
|
|||||||
-- Add CPU and memory resource limits per app
|
|
||||||
ALTER TABLE apps ADD COLUMN cpu_limit REAL;
|
|
||||||
ALTER TABLE apps ADD COLUMN memory_limit INTEGER;
|
|
||||||
@@ -1,41 +0,0 @@
|
|||||||
package database
|
|
||||||
|
|
||||||
import (
|
|
||||||
"log/slog"
|
|
||||||
"os"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
|
||||||
)
|
|
||||||
|
|
||||||
// NewTestDatabase creates an in-memory Database for testing.
|
|
||||||
// It runs migrations so all tables are available.
|
|
||||||
func NewTestDatabase(t *testing.T) *Database {
|
|
||||||
t.Helper()
|
|
||||||
|
|
||||||
tmpDir := t.TempDir()
|
|
||||||
|
|
||||||
cfg := &config.Config{
|
|
||||||
DataDir: tmpDir,
|
|
||||||
}
|
|
||||||
|
|
||||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
|
||||||
logWrapper := logger.NewForTest(log)
|
|
||||||
|
|
||||||
db, err := New(nil, Params{
|
|
||||||
Logger: logWrapper,
|
|
||||||
Config: cfg,
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to create test database: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
t.Cleanup(func() {
|
|
||||||
if db.database != nil {
|
|
||||||
_ = db.database.Close()
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
return db
|
|
||||||
}
|
|
||||||
@@ -14,7 +14,7 @@ import (
|
|||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
dockertypes "github.com/docker/docker/api/types"
|
"github.com/docker/docker/api/types"
|
||||||
"github.com/docker/docker/api/types/container"
|
"github.com/docker/docker/api/types/container"
|
||||||
"github.com/docker/docker/api/types/filters"
|
"github.com/docker/docker/api/types/filters"
|
||||||
"github.com/docker/docker/api/types/image"
|
"github.com/docker/docker/api/types/image"
|
||||||
@@ -25,9 +25,8 @@ import (
|
|||||||
"github.com/docker/go-connections/nat"
|
"github.com/docker/go-connections/nat"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// sshKeyPermissions is the file permission for SSH private keys.
|
// sshKeyPermissions is the file permission for SSH private keys.
|
||||||
@@ -117,7 +116,7 @@ type BuildImageOptions struct {
|
|||||||
func (c *Client) BuildImage(
|
func (c *Client) BuildImage(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
opts BuildImageOptions,
|
opts BuildImageOptions,
|
||||||
) (ImageID, error) {
|
) (string, error) {
|
||||||
if c.docker == nil {
|
if c.docker == nil {
|
||||||
return "", ErrNotConnected
|
return "", ErrNotConnected
|
||||||
}
|
}
|
||||||
@@ -145,8 +144,6 @@ type CreateContainerOptions struct {
|
|||||||
Volumes []VolumeMount
|
Volumes []VolumeMount
|
||||||
Ports []PortMapping
|
Ports []PortMapping
|
||||||
Network string
|
Network string
|
||||||
CPULimit float64 // CPU cores (e.g. 0.5 = half a core, 2.0 = two cores). 0 means unlimited.
|
|
||||||
MemoryLimit int64 // Memory in bytes. 0 means unlimited.
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// VolumeMount represents a volume mount.
|
// VolumeMount represents a volume mount.
|
||||||
@@ -163,14 +160,6 @@ type PortMapping struct {
|
|||||||
Protocol string // "tcp" or "udp"
|
Protocol string // "tcp" or "udp"
|
||||||
}
|
}
|
||||||
|
|
||||||
// nanoCPUsPerCPU is the number of NanoCPUs per CPU core.
|
|
||||||
const nanoCPUsPerCPU = 1e9
|
|
||||||
|
|
||||||
// cpuLimitToNanoCPUs converts a CPU limit (e.g. 0.5 cores) to Docker NanoCPUs.
|
|
||||||
func cpuLimitToNanoCPUs(cpuLimit float64) int64 {
|
|
||||||
return int64(cpuLimit * nanoCPUsPerCPU)
|
|
||||||
}
|
|
||||||
|
|
||||||
// buildPortConfig converts port mappings to Docker port configuration.
|
// buildPortConfig converts port mappings to Docker port configuration.
|
||||||
func buildPortConfig(ports []PortMapping) (nat.PortSet, nat.PortMap) {
|
func buildPortConfig(ports []PortMapping) (nat.PortSet, nat.PortMap) {
|
||||||
exposedPorts := make(nat.PortSet)
|
exposedPorts := make(nat.PortSet)
|
||||||
@@ -195,22 +184,28 @@ func buildPortConfig(ports []PortMapping) (nat.PortSet, nat.PortMap) {
|
|||||||
return exposedPorts, portBindings
|
return exposedPorts, portBindings
|
||||||
}
|
}
|
||||||
|
|
||||||
// buildEnvSlice converts an env map to a Docker-compatible env slice.
|
// CreateContainer creates a new container.
|
||||||
func buildEnvSlice(env map[string]string) []string {
|
func (c *Client) CreateContainer(
|
||||||
envSlice := make([]string, 0, len(env))
|
ctx context.Context,
|
||||||
|
opts CreateContainerOptions,
|
||||||
|
) (string, error) {
|
||||||
|
if c.docker == nil {
|
||||||
|
return "", ErrNotConnected
|
||||||
|
}
|
||||||
|
|
||||||
for key, val := range env {
|
c.log.Info("creating container", "name", opts.Name, "image", opts.Image)
|
||||||
|
|
||||||
|
// Convert env map to slice
|
||||||
|
envSlice := make([]string, 0, len(opts.Env))
|
||||||
|
|
||||||
|
for key, val := range opts.Env {
|
||||||
envSlice = append(envSlice, key+"="+val)
|
envSlice = append(envSlice, key+"="+val)
|
||||||
}
|
}
|
||||||
|
|
||||||
return envSlice
|
// Convert volumes to mounts
|
||||||
}
|
mounts := make([]mount.Mount, 0, len(opts.Volumes))
|
||||||
|
|
||||||
// buildMounts converts volume mounts to Docker mount configuration.
|
for _, vol := range opts.Volumes {
|
||||||
func buildMounts(volumes []VolumeMount) []mount.Mount {
|
|
||||||
mounts := make([]mount.Mount, 0, len(volumes))
|
|
||||||
|
|
||||||
for _, vol := range volumes {
|
|
||||||
mounts = append(mounts, mount.Mount{
|
mounts = append(mounts, mount.Mount{
|
||||||
Type: mount.TypeBind,
|
Type: mount.TypeBind,
|
||||||
Source: vol.HostPath,
|
Source: vol.HostPath,
|
||||||
@@ -219,49 +214,21 @@ func buildMounts(volumes []VolumeMount) []mount.Mount {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
return mounts
|
// Convert ports to exposed ports and port bindings
|
||||||
}
|
|
||||||
|
|
||||||
// buildResources builds Docker resource constraints from container options.
|
|
||||||
func buildResources(opts CreateContainerOptions) container.Resources {
|
|
||||||
resources := container.Resources{}
|
|
||||||
|
|
||||||
if opts.CPULimit > 0 {
|
|
||||||
resources.NanoCPUs = cpuLimitToNanoCPUs(opts.CPULimit)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.MemoryLimit > 0 {
|
|
||||||
resources.Memory = opts.MemoryLimit
|
|
||||||
}
|
|
||||||
|
|
||||||
return resources
|
|
||||||
}
|
|
||||||
|
|
||||||
// CreateContainer creates a new container.
|
|
||||||
func (c *Client) CreateContainer(
|
|
||||||
ctx context.Context,
|
|
||||||
opts CreateContainerOptions,
|
|
||||||
) (ContainerID, error) {
|
|
||||||
if c.docker == nil {
|
|
||||||
return "", ErrNotConnected
|
|
||||||
}
|
|
||||||
|
|
||||||
c.log.Info("creating container", "name", opts.Name, "image", opts.Image)
|
|
||||||
|
|
||||||
exposedPorts, portBindings := buildPortConfig(opts.Ports)
|
exposedPorts, portBindings := buildPortConfig(opts.Ports)
|
||||||
|
|
||||||
|
// Create container
|
||||||
resp, err := c.docker.ContainerCreate(ctx,
|
resp, err := c.docker.ContainerCreate(ctx,
|
||||||
&container.Config{
|
&container.Config{
|
||||||
Image: opts.Image,
|
Image: opts.Image,
|
||||||
Env: buildEnvSlice(opts.Env),
|
Env: envSlice,
|
||||||
Labels: opts.Labels,
|
Labels: opts.Labels,
|
||||||
ExposedPorts: exposedPorts,
|
ExposedPorts: exposedPorts,
|
||||||
},
|
},
|
||||||
&container.HostConfig{
|
&container.HostConfig{
|
||||||
Mounts: buildMounts(opts.Volumes),
|
Mounts: mounts,
|
||||||
PortBindings: portBindings,
|
PortBindings: portBindings,
|
||||||
NetworkMode: container.NetworkMode(opts.Network),
|
NetworkMode: container.NetworkMode(opts.Network),
|
||||||
Resources: buildResources(opts),
|
|
||||||
RestartPolicy: container.RestartPolicy{
|
RestartPolicy: container.RestartPolicy{
|
||||||
Name: container.RestartPolicyUnlessStopped,
|
Name: container.RestartPolicyUnlessStopped,
|
||||||
},
|
},
|
||||||
@@ -274,18 +241,18 @@ func (c *Client) CreateContainer(
|
|||||||
return "", fmt.Errorf("failed to create container: %w", err)
|
return "", fmt.Errorf("failed to create container: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
return ContainerID(resp.ID), nil
|
return resp.ID, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// StartContainer starts a container.
|
// StartContainer starts a container.
|
||||||
func (c *Client) StartContainer(ctx context.Context, containerID ContainerID) error {
|
func (c *Client) StartContainer(ctx context.Context, containerID string) error {
|
||||||
if c.docker == nil {
|
if c.docker == nil {
|
||||||
return ErrNotConnected
|
return ErrNotConnected
|
||||||
}
|
}
|
||||||
|
|
||||||
c.log.Info("starting container", "id", containerID)
|
c.log.Info("starting container", "id", containerID)
|
||||||
|
|
||||||
err := c.docker.ContainerStart(ctx, containerID.String(), container.StartOptions{})
|
err := c.docker.ContainerStart(ctx, containerID, container.StartOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to start container: %w", err)
|
return fmt.Errorf("failed to start container: %w", err)
|
||||||
}
|
}
|
||||||
@@ -294,7 +261,7 @@ func (c *Client) StartContainer(ctx context.Context, containerID ContainerID) er
|
|||||||
}
|
}
|
||||||
|
|
||||||
// StopContainer stops a container.
|
// StopContainer stops a container.
|
||||||
func (c *Client) StopContainer(ctx context.Context, containerID ContainerID) error {
|
func (c *Client) StopContainer(ctx context.Context, containerID string) error {
|
||||||
if c.docker == nil {
|
if c.docker == nil {
|
||||||
return ErrNotConnected
|
return ErrNotConnected
|
||||||
}
|
}
|
||||||
@@ -303,7 +270,7 @@ func (c *Client) StopContainer(ctx context.Context, containerID ContainerID) err
|
|||||||
|
|
||||||
timeout := stopTimeoutSeconds
|
timeout := stopTimeoutSeconds
|
||||||
|
|
||||||
err := c.docker.ContainerStop(ctx, containerID.String(), container.StopOptions{Timeout: &timeout})
|
err := c.docker.ContainerStop(ctx, containerID, container.StopOptions{Timeout: &timeout})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to stop container: %w", err)
|
return fmt.Errorf("failed to stop container: %w", err)
|
||||||
}
|
}
|
||||||
@@ -314,7 +281,7 @@ func (c *Client) StopContainer(ctx context.Context, containerID ContainerID) err
|
|||||||
// RemoveContainer removes a container.
|
// RemoveContainer removes a container.
|
||||||
func (c *Client) RemoveContainer(
|
func (c *Client) RemoveContainer(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
containerID ContainerID,
|
containerID string,
|
||||||
force bool,
|
force bool,
|
||||||
) error {
|
) error {
|
||||||
if c.docker == nil {
|
if c.docker == nil {
|
||||||
@@ -323,7 +290,7 @@ func (c *Client) RemoveContainer(
|
|||||||
|
|
||||||
c.log.Info("removing container", "id", containerID, "force", force)
|
c.log.Info("removing container", "id", containerID, "force", force)
|
||||||
|
|
||||||
err := c.docker.ContainerRemove(ctx, containerID.String(), container.RemoveOptions{Force: force})
|
err := c.docker.ContainerRemove(ctx, containerID, container.RemoveOptions{Force: force})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to remove container: %w", err)
|
return fmt.Errorf("failed to remove container: %w", err)
|
||||||
}
|
}
|
||||||
@@ -334,7 +301,7 @@ func (c *Client) RemoveContainer(
|
|||||||
// ContainerLogs returns the logs for a container.
|
// ContainerLogs returns the logs for a container.
|
||||||
func (c *Client) ContainerLogs(
|
func (c *Client) ContainerLogs(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
containerID ContainerID,
|
containerID string,
|
||||||
tail string,
|
tail string,
|
||||||
) (string, error) {
|
) (string, error) {
|
||||||
if c.docker == nil {
|
if c.docker == nil {
|
||||||
@@ -347,7 +314,7 @@ func (c *Client) ContainerLogs(
|
|||||||
Tail: tail,
|
Tail: tail,
|
||||||
}
|
}
|
||||||
|
|
||||||
reader, err := c.docker.ContainerLogs(ctx, containerID.String(), opts)
|
reader, err := c.docker.ContainerLogs(ctx, containerID, opts)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", fmt.Errorf("failed to get container logs: %w", err)
|
return "", fmt.Errorf("failed to get container logs: %w", err)
|
||||||
}
|
}
|
||||||
@@ -370,13 +337,13 @@ func (c *Client) ContainerLogs(
|
|||||||
// IsContainerRunning checks if a container is running.
|
// IsContainerRunning checks if a container is running.
|
||||||
func (c *Client) IsContainerRunning(
|
func (c *Client) IsContainerRunning(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
containerID ContainerID,
|
containerID string,
|
||||||
) (bool, error) {
|
) (bool, error) {
|
||||||
if c.docker == nil {
|
if c.docker == nil {
|
||||||
return false, ErrNotConnected
|
return false, ErrNotConnected
|
||||||
}
|
}
|
||||||
|
|
||||||
inspect, err := c.docker.ContainerInspect(ctx, containerID.String())
|
inspect, err := c.docker.ContainerInspect(ctx, containerID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false, fmt.Errorf("failed to inspect container: %w", err)
|
return false, fmt.Errorf("failed to inspect container: %w", err)
|
||||||
}
|
}
|
||||||
@@ -387,13 +354,13 @@ func (c *Client) IsContainerRunning(
|
|||||||
// IsContainerHealthy checks if a container is healthy.
|
// IsContainerHealthy checks if a container is healthy.
|
||||||
func (c *Client) IsContainerHealthy(
|
func (c *Client) IsContainerHealthy(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
containerID ContainerID,
|
containerID string,
|
||||||
) (bool, error) {
|
) (bool, error) {
|
||||||
if c.docker == nil {
|
if c.docker == nil {
|
||||||
return false, ErrNotConnected
|
return false, ErrNotConnected
|
||||||
}
|
}
|
||||||
|
|
||||||
inspect, err := c.docker.ContainerInspect(ctx, containerID.String())
|
inspect, err := c.docker.ContainerInspect(ctx, containerID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false, fmt.Errorf("failed to inspect container: %w", err)
|
return false, fmt.Errorf("failed to inspect container: %w", err)
|
||||||
}
|
}
|
||||||
@@ -411,7 +378,7 @@ const LabelUpaasID = "upaas.id"
|
|||||||
|
|
||||||
// ContainerInfo contains basic information about a container.
|
// ContainerInfo contains basic information about a container.
|
||||||
type ContainerInfo struct {
|
type ContainerInfo struct {
|
||||||
ID ContainerID
|
ID string
|
||||||
Running bool
|
Running bool
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -446,7 +413,7 @@ func (c *Client) FindContainerByAppID(
|
|||||||
ctr := containers[0]
|
ctr := containers[0]
|
||||||
|
|
||||||
return &ContainerInfo{
|
return &ContainerInfo{
|
||||||
ID: ContainerID(ctr.ID),
|
ID: ctr.ID,
|
||||||
Running: ctr.State == "running",
|
Running: ctr.State == "running",
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
@@ -515,8 +482,8 @@ func (c *Client) CloneRepo(
|
|||||||
|
|
||||||
// RemoveImage removes a Docker image by ID or tag.
|
// RemoveImage removes a Docker image by ID or tag.
|
||||||
// It returns nil if the image was successfully removed or does not exist.
|
// It returns nil if the image was successfully removed or does not exist.
|
||||||
func (c *Client) RemoveImage(ctx context.Context, imageID ImageID) error {
|
func (c *Client) RemoveImage(ctx context.Context, imageID string) error {
|
||||||
_, err := c.docker.ImageRemove(ctx, imageID.String(), image.RemoveOptions{
|
_, err := c.docker.ImageRemove(ctx, imageID, image.RemoveOptions{
|
||||||
Force: true,
|
Force: true,
|
||||||
PruneChildren: true,
|
PruneChildren: true,
|
||||||
})
|
})
|
||||||
@@ -530,7 +497,7 @@ func (c *Client) RemoveImage(ctx context.Context, imageID ImageID) error {
|
|||||||
func (c *Client) performBuild(
|
func (c *Client) performBuild(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
opts BuildImageOptions,
|
opts BuildImageOptions,
|
||||||
) (ImageID, error) {
|
) (string, error) {
|
||||||
// Create tar archive of build context
|
// Create tar archive of build context
|
||||||
tarArchive, err := archive.TarWithOptions(opts.ContextDir, &archive.TarOptions{})
|
tarArchive, err := archive.TarWithOptions(opts.ContextDir, &archive.TarOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -545,7 +512,7 @@ func (c *Client) performBuild(
|
|||||||
}()
|
}()
|
||||||
|
|
||||||
// Build image
|
// Build image
|
||||||
resp, err := c.docker.ImageBuild(ctx, tarArchive, dockertypes.ImageBuildOptions{
|
resp, err := c.docker.ImageBuild(ctx, tarArchive, types.ImageBuildOptions{
|
||||||
Dockerfile: opts.DockerfilePath,
|
Dockerfile: opts.DockerfilePath,
|
||||||
Tags: opts.Tags,
|
Tags: opts.Tags,
|
||||||
Remove: true,
|
Remove: true,
|
||||||
@@ -575,7 +542,7 @@ func (c *Client) performBuild(
|
|||||||
return "", fmt.Errorf("failed to inspect image: %w", inspectErr)
|
return "", fmt.Errorf("failed to inspect image: %w", inspectErr)
|
||||||
}
|
}
|
||||||
|
|
||||||
return ImageID(inspect.ID), nil
|
return inspect.ID, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return "", nil
|
return "", nil
|
||||||
@@ -636,22 +603,22 @@ func (c *Client) performClone(ctx context.Context, cfg *cloneConfig) (*CloneResu
|
|||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
gitContainerID, err := c.createGitContainer(ctx, cfg)
|
containerID, err := c.createGitContainer(ctx, cfg)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
defer func() {
|
defer func() {
|
||||||
_ = c.docker.ContainerRemove(ctx, gitContainerID.String(), container.RemoveOptions{Force: true})
|
_ = c.docker.ContainerRemove(ctx, containerID, container.RemoveOptions{Force: true})
|
||||||
}()
|
}()
|
||||||
|
|
||||||
return c.runGitClone(ctx, gitContainerID)
|
return c.runGitClone(ctx, containerID)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Client) createGitContainer(
|
func (c *Client) createGitContainer(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
cfg *cloneConfig,
|
cfg *cloneConfig,
|
||||||
) (ContainerID, error) {
|
) (string, error) {
|
||||||
gitSSHCmd := "ssh -i /keys/deploy_key -o StrictHostKeyChecking=no"
|
gitSSHCmd := "ssh -i /keys/deploy_key -o StrictHostKeyChecking=no"
|
||||||
|
|
||||||
// Build the git command using environment variables to avoid shell injection.
|
// Build the git command using environment variables to avoid shell injection.
|
||||||
@@ -708,16 +675,16 @@ func (c *Client) createGitContainer(
|
|||||||
return "", fmt.Errorf("failed to create git container: %w", err)
|
return "", fmt.Errorf("failed to create git container: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
return ContainerID(resp.ID), nil
|
return resp.ID, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Client) runGitClone(ctx context.Context, containerID ContainerID) (*CloneResult, error) {
|
func (c *Client) runGitClone(ctx context.Context, containerID string) (*CloneResult, error) {
|
||||||
err := c.docker.ContainerStart(ctx, containerID.String(), container.StartOptions{})
|
err := c.docker.ContainerStart(ctx, containerID, container.StartOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("failed to start git container: %w", err)
|
return nil, fmt.Errorf("failed to start git container: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
statusCh, errCh := c.docker.ContainerWait(ctx, containerID.String(), container.WaitConditionNotRunning)
|
statusCh, errCh := c.docker.ContainerWait(ctx, containerID, container.WaitConditionNotRunning)
|
||||||
|
|
||||||
select {
|
select {
|
||||||
case err := <-errCh:
|
case err := <-errCh:
|
||||||
|
|||||||
@@ -1,31 +0,0 @@
|
|||||||
package docker //nolint:testpackage // tests unexported cpuLimitToNanoCPUs
|
|
||||||
|
|
||||||
import (
|
|
||||||
"testing"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestCpuLimitToNanoCPUs(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
cpuLimit float64
|
|
||||||
expected int64
|
|
||||||
}{
|
|
||||||
{"one core", 1.0, 1_000_000_000},
|
|
||||||
{"half core", 0.5, 500_000_000},
|
|
||||||
{"two cores", 2.0, 2_000_000_000},
|
|
||||||
{"quarter core", 0.25, 250_000_000},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
got := cpuLimitToNanoCPUs(tt.cpuLimit)
|
|
||||||
if got != tt.expected {
|
|
||||||
t.Errorf("cpuLimitToNanoCPUs(%v) = %d, want %d", tt.cpuLimit, got, tt.expected)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,13 +0,0 @@
|
|||||||
package docker
|
|
||||||
|
|
||||||
// ImageID is a Docker image identifier (ID or tag).
|
|
||||||
type ImageID string
|
|
||||||
|
|
||||||
// String implements the fmt.Stringer interface.
|
|
||||||
func (id ImageID) String() string { return string(id) }
|
|
||||||
|
|
||||||
// ContainerID is a Docker container identifier.
|
|
||||||
type ContainerID string
|
|
||||||
|
|
||||||
// String implements the fmt.Stringer interface.
|
|
||||||
func (id ContainerID) String() string { return string(id) }
|
|
||||||
@@ -7,7 +7,7 @@ import (
|
|||||||
|
|
||||||
"github.com/go-chi/chi/v5"
|
"github.com/go-chi/chi/v5"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
)
|
)
|
||||||
|
|
||||||
// apiAppResponse is the JSON representation of an app.
|
// apiAppResponse is the JSON representation of an app.
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ import (
|
|||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/service/app"
|
"git.eeqj.de/sneak/upaas/internal/service/app"
|
||||||
)
|
)
|
||||||
|
|
||||||
// apiRouter builds a chi router with the API routes using session auth middleware.
|
// apiRouter builds a chi router with the API routes using session auth middleware.
|
||||||
|
|||||||
@@ -15,9 +15,9 @@ import (
|
|||||||
|
|
||||||
"github.com/go-chi/chi/v5"
|
"github.com/go-chi/chi/v5"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
"sneak.berlin/go/upaas/internal/service/app"
|
"git.eeqj.de/sneak/upaas/internal/service/app"
|
||||||
"sneak.berlin/go/upaas/templates"
|
"git.eeqj.de/sneak/upaas/templates"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -54,18 +54,12 @@ func (h *Handlers) HandleAppCreate() http.HandlerFunc { //nolint:funlen // valid
|
|||||||
repoURL := request.FormValue("repo_url")
|
repoURL := request.FormValue("repo_url")
|
||||||
branch := request.FormValue("branch")
|
branch := request.FormValue("branch")
|
||||||
dockerfilePath := request.FormValue("dockerfile_path")
|
dockerfilePath := request.FormValue("dockerfile_path")
|
||||||
dockerNetwork := request.FormValue("docker_network")
|
|
||||||
ntfyTopic := request.FormValue("ntfy_topic")
|
|
||||||
slackWebhook := request.FormValue("slack_webhook")
|
|
||||||
|
|
||||||
data := h.addGlobals(map[string]any{
|
data := h.addGlobals(map[string]any{
|
||||||
"Name": name,
|
"Name": name,
|
||||||
"RepoURL": repoURL,
|
"RepoURL": repoURL,
|
||||||
"Branch": branch,
|
"Branch": branch,
|
||||||
"DockerfilePath": dockerfilePath,
|
"DockerfilePath": dockerfilePath,
|
||||||
"DockerNetwork": dockerNetwork,
|
|
||||||
"NtfyTopic": ntfyTopic,
|
|
||||||
"SlackWebhook": slackWebhook,
|
|
||||||
}, request)
|
}, request)
|
||||||
|
|
||||||
if name == "" || repoURL == "" {
|
if name == "" || repoURL == "" {
|
||||||
@@ -106,9 +100,6 @@ func (h *Handlers) HandleAppCreate() http.HandlerFunc { //nolint:funlen // valid
|
|||||||
RepoURL: repoURL,
|
RepoURL: repoURL,
|
||||||
Branch: branch,
|
Branch: branch,
|
||||||
DockerfilePath: dockerfilePath,
|
DockerfilePath: dockerfilePath,
|
||||||
DockerNetwork: dockerNetwork,
|
|
||||||
NtfyTopic: ntfyTopic,
|
|
||||||
SlackWebhook: slackWebhook,
|
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
if createErr != nil {
|
if createErr != nil {
|
||||||
@@ -257,19 +248,23 @@ func (h *Handlers) HandleAppUpdate() http.HandlerFunc { //nolint:funlen // valid
|
|||||||
application.RepoURL = request.FormValue("repo_url")
|
application.RepoURL = request.FormValue("repo_url")
|
||||||
application.Branch = request.FormValue("branch")
|
application.Branch = request.FormValue("branch")
|
||||||
application.DockerfilePath = request.FormValue("dockerfile_path")
|
application.DockerfilePath = request.FormValue("dockerfile_path")
|
||||||
application.DockerNetwork = optionalNullString(request.FormValue("docker_network"))
|
|
||||||
application.NtfyTopic = optionalNullString(request.FormValue("ntfy_topic"))
|
|
||||||
application.SlackWebhook = optionalNullString(request.FormValue("slack_webhook"))
|
|
||||||
|
|
||||||
limitsErr := applyResourceLimits(application, request)
|
if network := request.FormValue("docker_network"); network != "" {
|
||||||
if limitsErr != "" {
|
application.DockerNetwork = sql.NullString{String: network, Valid: true}
|
||||||
data := h.addGlobals(map[string]any{
|
} else {
|
||||||
"App": application,
|
application.DockerNetwork = sql.NullString{}
|
||||||
"Error": limitsErr,
|
}
|
||||||
}, request)
|
|
||||||
h.renderTemplate(writer, tmpl, "app_edit.html", data)
|
|
||||||
|
|
||||||
return
|
if ntfy := request.FormValue("ntfy_topic"); ntfy != "" {
|
||||||
|
application.NtfyTopic = sql.NullString{String: ntfy, Valid: true}
|
||||||
|
} else {
|
||||||
|
application.NtfyTopic = sql.NullString{}
|
||||||
|
}
|
||||||
|
|
||||||
|
if slack := request.FormValue("slack_webhook"); slack != "" {
|
||||||
|
application.SlackWebhook = sql.NullString{String: slack, Valid: true}
|
||||||
|
} else {
|
||||||
|
application.SlackWebhook = sql.NullString{}
|
||||||
}
|
}
|
||||||
|
|
||||||
saveErr := application.Save(request.Context())
|
saveErr := application.Save(request.Context())
|
||||||
@@ -899,92 +894,50 @@ func (h *Handlers) addKeyValueToApp(
|
|||||||
http.Redirect(writer, request, "/apps/"+application.ID, http.StatusSeeOther)
|
http.Redirect(writer, request, "/apps/"+application.ID, http.StatusSeeOther)
|
||||||
}
|
}
|
||||||
|
|
||||||
// envPairJSON represents a key-value pair in the JSON request body.
|
// HandleEnvVarAdd handles adding an environment variable.
|
||||||
type envPairJSON struct {
|
func (h *Handlers) HandleEnvVarAdd() http.HandlerFunc {
|
||||||
Key string `json:"key"`
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
Value string `json:"value"`
|
h.addKeyValueToApp(
|
||||||
|
writer,
|
||||||
|
request,
|
||||||
|
func(ctx context.Context, application *models.App, key, value string) error {
|
||||||
|
envVar := models.NewEnvVar(h.db)
|
||||||
|
envVar.AppID = application.ID
|
||||||
|
envVar.Key = key
|
||||||
|
envVar.Value = value
|
||||||
|
|
||||||
|
return envVar.Save(ctx)
|
||||||
|
},
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// envVarMaxBodyBytes is the maximum allowed request body size for env var saves (1 MB).
|
// HandleEnvVarDelete handles deleting an environment variable.
|
||||||
const envVarMaxBodyBytes = 1 << 20
|
func (h *Handlers) HandleEnvVarDelete() http.HandlerFunc {
|
||||||
|
|
||||||
// validateEnvPairs validates env var pairs.
|
|
||||||
// It rejects empty keys and duplicate keys (returns a non-empty error string).
|
|
||||||
func validateEnvPairs(pairs []envPairJSON) ([]models.EnvVarPair, string) {
|
|
||||||
seen := make(map[string]bool, len(pairs))
|
|
||||||
|
|
||||||
result := make([]models.EnvVarPair, 0, len(pairs))
|
|
||||||
|
|
||||||
for _, p := range pairs {
|
|
||||||
trimmedKey := strings.TrimSpace(p.Key)
|
|
||||||
if trimmedKey == "" {
|
|
||||||
return nil, "empty environment variable key is not allowed"
|
|
||||||
}
|
|
||||||
|
|
||||||
if seen[trimmedKey] {
|
|
||||||
return nil, "duplicate environment variable key: " + trimmedKey
|
|
||||||
}
|
|
||||||
|
|
||||||
seen[trimmedKey] = true
|
|
||||||
|
|
||||||
result = append(result, models.EnvVarPair{Key: trimmedKey, Value: p.Value})
|
|
||||||
}
|
|
||||||
|
|
||||||
return result, ""
|
|
||||||
}
|
|
||||||
|
|
||||||
// HandleEnvVarSave handles bulk saving of all environment variables.
|
|
||||||
// It reads a JSON array of {key, value} objects from the request body,
|
|
||||||
// deletes all existing env vars for the app, and inserts the full
|
|
||||||
// submitted set atomically within a database transaction.
|
|
||||||
// Duplicate keys are rejected with a 400 Bad Request error.
|
|
||||||
func (h *Handlers) HandleEnvVarSave() http.HandlerFunc {
|
|
||||||
return func(writer http.ResponseWriter, request *http.Request) {
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
appID := chi.URLParam(request, "id")
|
appID := chi.URLParam(request, "id")
|
||||||
|
envVarIDStr := chi.URLParam(request, "varID")
|
||||||
|
|
||||||
application, findErr := models.FindApp(request.Context(), h.db, appID)
|
envVarID, parseErr := strconv.ParseInt(envVarIDStr, 10, 64)
|
||||||
if findErr != nil || application == nil {
|
if parseErr != nil {
|
||||||
http.NotFound(writer, request)
|
http.NotFound(writer, request)
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Limit request body size to prevent abuse
|
envVar, findErr := models.FindEnvVar(request.Context(), h.db, envVarID)
|
||||||
request.Body = http.MaxBytesReader(writer, request.Body, envVarMaxBodyBytes)
|
if findErr != nil || envVar == nil || envVar.AppID != appID {
|
||||||
|
http.NotFound(writer, request)
|
||||||
var pairs []envPairJSON
|
|
||||||
|
|
||||||
decodeErr := json.NewDecoder(request.Body).Decode(&pairs)
|
|
||||||
if decodeErr != nil {
|
|
||||||
h.respondJSON(writer, request, map[string]string{
|
|
||||||
"error": "invalid request body",
|
|
||||||
}, http.StatusBadRequest)
|
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
modelPairs, validationErr := validateEnvPairs(pairs)
|
deleteErr := envVar.Delete(request.Context())
|
||||||
if validationErr != "" {
|
if deleteErr != nil {
|
||||||
h.respondJSON(writer, request, map[string]string{
|
h.log.Error("failed to delete env var", "error", deleteErr)
|
||||||
"error": validationErr,
|
|
||||||
}, http.StatusBadRequest)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
}
|
||||||
|
|
||||||
replaceErr := models.ReplaceEnvVarsByAppID(
|
http.Redirect(writer, request, "/apps/"+appID, http.StatusSeeOther)
|
||||||
request.Context(), h.db, application.ID, modelPairs,
|
|
||||||
)
|
|
||||||
if replaceErr != nil {
|
|
||||||
h.log.Error("failed to replace env vars", "error", replaceErr)
|
|
||||||
h.respondJSON(writer, request, map[string]string{
|
|
||||||
"error": "failed to save environment variables",
|
|
||||||
}, http.StatusInternalServerError)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
h.respondJSON(writer, request, map[string]bool{"ok": true}, http.StatusOK)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1243,6 +1196,59 @@ func ValidateVolumePath(p string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// HandleEnvVarEdit handles editing an existing environment variable.
|
||||||
|
func (h *Handlers) HandleEnvVarEdit() http.HandlerFunc {
|
||||||
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
appID := chi.URLParam(request, "id")
|
||||||
|
envVarIDStr := chi.URLParam(request, "varID")
|
||||||
|
|
||||||
|
envVarID, parseErr := strconv.ParseInt(envVarIDStr, 10, 64)
|
||||||
|
if parseErr != nil {
|
||||||
|
http.NotFound(writer, request)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
envVar, findErr := models.FindEnvVar(request.Context(), h.db, envVarID)
|
||||||
|
if findErr != nil || envVar == nil || envVar.AppID != appID {
|
||||||
|
http.NotFound(writer, request)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
formErr := request.ParseForm()
|
||||||
|
if formErr != nil {
|
||||||
|
http.Error(writer, "Bad Request", http.StatusBadRequest)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
key := request.FormValue("key")
|
||||||
|
value := request.FormValue("value")
|
||||||
|
|
||||||
|
if key == "" || value == "" {
|
||||||
|
http.Redirect(writer, request, "/apps/"+appID, http.StatusSeeOther)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
envVar.Key = key
|
||||||
|
envVar.Value = value
|
||||||
|
|
||||||
|
saveErr := envVar.Save(request.Context())
|
||||||
|
if saveErr != nil {
|
||||||
|
h.log.Error("failed to update env var", "error", saveErr)
|
||||||
|
}
|
||||||
|
|
||||||
|
http.Redirect(
|
||||||
|
writer,
|
||||||
|
request,
|
||||||
|
"/apps/"+appID+"?success=env-updated",
|
||||||
|
http.StatusSeeOther,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// HandleLabelEdit handles editing an existing label.
|
// HandleLabelEdit handles editing an existing label.
|
||||||
func (h *Handlers) HandleLabelEdit() http.HandlerFunc {
|
func (h *Handlers) HandleLabelEdit() http.HandlerFunc {
|
||||||
return func(writer http.ResponseWriter, request *http.Request) {
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
@@ -1364,129 +1370,6 @@ func validateVolumePaths(hostPath, containerPath string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// ErrInvalidMemoryFormat is returned when a memory limit string cannot be parsed.
|
|
||||||
var ErrInvalidMemoryFormat = errors.New(
|
|
||||||
"must be a number with optional unit suffix (e.g. 256m, 1g, 512000000)",
|
|
||||||
)
|
|
||||||
|
|
||||||
// ErrNegativeValue is returned when a resource limit is negative.
|
|
||||||
var ErrNegativeValue = errors.New("value must be positive")
|
|
||||||
|
|
||||||
// Memory unit byte multipliers.
|
|
||||||
const (
|
|
||||||
kilobyte = 1024
|
|
||||||
megabyte = 1024 * 1024
|
|
||||||
gigabyte = 1024 * 1024 * 1024
|
|
||||||
)
|
|
||||||
|
|
||||||
// optionalNullString converts a form value to a sql.NullString.
|
|
||||||
// Returns a valid NullString if non-empty, invalid (NULL) if empty.
|
|
||||||
func optionalNullString(s string) sql.NullString {
|
|
||||||
if s != "" {
|
|
||||||
return sql.NullString{String: s, Valid: true}
|
|
||||||
}
|
|
||||||
|
|
||||||
return sql.NullString{}
|
|
||||||
}
|
|
||||||
|
|
||||||
// applyResourceLimits parses CPU and memory limit form values and applies them to the app.
|
|
||||||
// Returns an error message string if validation fails, or empty string on success.
|
|
||||||
func applyResourceLimits(application *models.App, request *http.Request) string {
|
|
||||||
cpuLimit, cpuErr := parseOptionalFloat64(request.FormValue("cpu_limit"))
|
|
||||||
if cpuErr != nil {
|
|
||||||
return "Invalid CPU limit: must be a positive number (e.g. 0.5, 1, 2)"
|
|
||||||
}
|
|
||||||
|
|
||||||
application.CPULimit = cpuLimit
|
|
||||||
|
|
||||||
memoryLimit, memErr := parseOptionalMemoryBytes(request.FormValue("memory_limit"))
|
|
||||||
if memErr != nil {
|
|
||||||
return "Invalid memory limit: " + memErr.Error()
|
|
||||||
}
|
|
||||||
|
|
||||||
application.MemoryLimit = memoryLimit
|
|
||||||
|
|
||||||
return ""
|
|
||||||
}
|
|
||||||
|
|
||||||
// memoryUnitMultiplier returns the byte multiplier for a memory unit suffix.
|
|
||||||
// Returns 0 if the suffix is not recognized.
|
|
||||||
func memoryUnitMultiplier(suffix byte) int64 {
|
|
||||||
switch suffix {
|
|
||||||
case 'k':
|
|
||||||
return kilobyte
|
|
||||||
case 'm':
|
|
||||||
return megabyte
|
|
||||||
case 'g':
|
|
||||||
return gigabyte
|
|
||||||
default:
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// parseOptionalFloat64 parses an optional float64 form field.
|
|
||||||
// Returns a valid NullFloat64 if the string is non-empty and parses to a positive number.
|
|
||||||
// Returns an empty NullFloat64 if the string is empty.
|
|
||||||
// Returns an error if the string is non-empty but invalid or non-positive.
|
|
||||||
func parseOptionalFloat64(s string) (sql.NullFloat64, error) {
|
|
||||||
s = strings.TrimSpace(s)
|
|
||||||
if s == "" {
|
|
||||||
return sql.NullFloat64{}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
val, err := strconv.ParseFloat(s, 64)
|
|
||||||
if err != nil {
|
|
||||||
return sql.NullFloat64{}, fmt.Errorf("invalid number: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if val <= 0 {
|
|
||||||
return sql.NullFloat64{}, ErrNegativeValue
|
|
||||||
}
|
|
||||||
|
|
||||||
return sql.NullFloat64{Float64: val, Valid: true}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// parseOptionalMemoryBytes parses an optional memory limit string into bytes.
|
|
||||||
// Accepts plain bytes (e.g. "536870912") or suffixed values (e.g. "512m", "1g", "256k").
|
|
||||||
// Returns a valid NullInt64 with bytes if non-empty, empty NullInt64 if blank.
|
|
||||||
func parseOptionalMemoryBytes(s string) (sql.NullInt64, error) {
|
|
||||||
s = strings.TrimSpace(s)
|
|
||||||
if s == "" {
|
|
||||||
return sql.NullInt64{}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
s = strings.ToLower(s)
|
|
||||||
|
|
||||||
// Check for unit suffix
|
|
||||||
multiplier := memoryUnitMultiplier(s[len(s)-1])
|
|
||||||
if multiplier > 0 {
|
|
||||||
numStr := s[:len(s)-1]
|
|
||||||
|
|
||||||
val, err := strconv.ParseFloat(numStr, 64)
|
|
||||||
if err != nil {
|
|
||||||
return sql.NullInt64{}, ErrInvalidMemoryFormat
|
|
||||||
}
|
|
||||||
|
|
||||||
if val <= 0 {
|
|
||||||
return sql.NullInt64{}, ErrNegativeValue
|
|
||||||
}
|
|
||||||
|
|
||||||
return sql.NullInt64{Int64: int64(val * float64(multiplier)), Valid: true}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Plain bytes
|
|
||||||
val, err := strconv.ParseInt(s, 10, 64)
|
|
||||||
if err != nil {
|
|
||||||
return sql.NullInt64{}, ErrInvalidMemoryFormat
|
|
||||||
}
|
|
||||||
|
|
||||||
if val <= 0 {
|
|
||||||
return sql.NullInt64{}, ErrNegativeValue
|
|
||||||
}
|
|
||||||
|
|
||||||
return sql.NullInt64{Int64: val, Valid: true}, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// formatDeployKey formats an SSH public key with a descriptive comment.
|
// formatDeployKey formats an SSH public key with a descriptive comment.
|
||||||
// Format: ssh-ed25519 AAAA... upaas_2025-01-15_myapp
|
// Format: ssh-ed25519 AAAA... upaas_2025-01-15_myapp
|
||||||
func formatDeployKey(pubKey string, createdAt time.Time, appName string) string {
|
func formatDeployKey(pubKey string, createdAt time.Time, appName string) string {
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package handlers
|
|||||||
import (
|
import (
|
||||||
"net/http"
|
"net/http"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/templates"
|
"git.eeqj.de/sneak/upaas/templates"
|
||||||
)
|
)
|
||||||
|
|
||||||
// HandleLoginGET returns the login page handler.
|
// HandleLoginGET returns the login page handler.
|
||||||
|
|||||||
@@ -4,8 +4,8 @@ import (
|
|||||||
"net/http"
|
"net/http"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
"sneak.berlin/go/upaas/templates"
|
"git.eeqj.de/sneak/upaas/templates"
|
||||||
)
|
)
|
||||||
|
|
||||||
// AppStats holds deployment statistics for an app.
|
// AppStats holds deployment statistics for an app.
|
||||||
|
|||||||
@@ -10,16 +10,16 @@ import (
|
|||||||
"github.com/gorilla/csrf"
|
"github.com/gorilla/csrf"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/docker"
|
"git.eeqj.de/sneak/upaas/internal/docker"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/healthcheck"
|
"git.eeqj.de/sneak/upaas/internal/healthcheck"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/service/app"
|
"git.eeqj.de/sneak/upaas/internal/service/app"
|
||||||
"sneak.berlin/go/upaas/internal/service/auth"
|
"git.eeqj.de/sneak/upaas/internal/service/auth"
|
||||||
"sneak.berlin/go/upaas/internal/service/deploy"
|
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
||||||
"sneak.berlin/go/upaas/internal/service/webhook"
|
"git.eeqj.de/sneak/upaas/internal/service/webhook"
|
||||||
"sneak.berlin/go/upaas/templates"
|
"git.eeqj.de/sneak/upaas/templates"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Params contains dependencies for Handlers.
|
// Params contains dependencies for Handlers.
|
||||||
|
|||||||
@@ -15,21 +15,21 @@ import (
|
|||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/docker"
|
"git.eeqj.de/sneak/upaas/internal/docker"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/handlers"
|
"git.eeqj.de/sneak/upaas/internal/handlers"
|
||||||
"sneak.berlin/go/upaas/internal/healthcheck"
|
"git.eeqj.de/sneak/upaas/internal/healthcheck"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/middleware"
|
"git.eeqj.de/sneak/upaas/internal/middleware"
|
||||||
"sneak.berlin/go/upaas/internal/service/app"
|
"git.eeqj.de/sneak/upaas/internal/service/app"
|
||||||
"sneak.berlin/go/upaas/internal/service/auth"
|
"git.eeqj.de/sneak/upaas/internal/service/auth"
|
||||||
"sneak.berlin/go/upaas/internal/service/deploy"
|
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
||||||
"sneak.berlin/go/upaas/internal/service/notify"
|
"git.eeqj.de/sneak/upaas/internal/service/notify"
|
||||||
"sneak.berlin/go/upaas/internal/service/webhook"
|
"git.eeqj.de/sneak/upaas/internal/service/webhook"
|
||||||
)
|
)
|
||||||
|
|
||||||
type testContext struct {
|
type testContext struct {
|
||||||
@@ -404,25 +404,6 @@ func TestHandleDashboard(t *testing.T) {
|
|||||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
assert.Equal(t, http.StatusOK, recorder.Code)
|
||||||
assert.Contains(t, recorder.Body.String(), "Applications")
|
assert.Contains(t, recorder.Body.String(), "Applications")
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("renders dashboard with apps without crashing on CSRFField", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
|
|
||||||
// Create an app so the template iterates over AppStats and hits .CSRFField
|
|
||||||
createTestApp(t, testCtx, "csrf-test-app")
|
|
||||||
|
|
||||||
request := httptest.NewRequest(http.MethodGet, "/", nil)
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
|
|
||||||
handler := testCtx.handlers.HandleDashboard()
|
|
||||||
handler.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusOK, recorder.Code,
|
|
||||||
"dashboard should not 500 when apps exist (CSRFField must be accessible)")
|
|
||||||
assert.Contains(t, recorder.Body.String(), "csrf-test-app")
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestHandleAppNew(t *testing.T) {
|
func TestHandleAppNew(t *testing.T) {
|
||||||
@@ -560,242 +541,45 @@ func testOwnershipVerification(t *testing.T, cfg ownedResourceTestConfig) {
|
|||||||
cfg.verifyFn(t, testCtx, resourceID)
|
cfg.verifyFn(t, testCtx, resourceID)
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestHandleEnvVarSaveBulk tests that HandleEnvVarSave replaces all env vars
|
// TestDeleteEnvVarOwnershipVerification tests that deleting an env var
|
||||||
// for an app with the submitted set (monolithic delete-all + insert-all).
|
// via another app's URL path returns 404 (IDOR prevention).
|
||||||
func TestHandleEnvVarSaveBulk(t *testing.T) {
|
func TestDeleteEnvVarOwnershipVerification(t *testing.T) { //nolint:dupl // intentionally similar IDOR test pattern
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
testOwnershipVerification(t, ownedResourceTestConfig{
|
||||||
createdApp := createTestApp(t, testCtx, "envvar-bulk-app")
|
appPrefix1: "envvar-owner-app",
|
||||||
|
appPrefix2: "envvar-other-app",
|
||||||
|
createFn: func(t *testing.T, tc *testContext, ownerApp *models.App) int64 {
|
||||||
|
t.Helper()
|
||||||
|
|
||||||
// Create some pre-existing env vars
|
envVar := models.NewEnvVar(tc.database)
|
||||||
for _, kv := range [][2]string{{"OLD_KEY", "old_value"}, {"REMOVE_ME", "gone"}} {
|
envVar.AppID = ownerApp.ID
|
||||||
ev := models.NewEnvVar(testCtx.database)
|
envVar.Key = "SECRET"
|
||||||
ev.AppID = createdApp.ID
|
envVar.Value = "hunter2"
|
||||||
ev.Key = kv[0]
|
require.NoError(t, envVar.Save(context.Background()))
|
||||||
ev.Value = kv[1]
|
|
||||||
require.NoError(t, ev.Save(context.Background()))
|
|
||||||
}
|
|
||||||
|
|
||||||
// Submit a new set as a JSON array of key/value objects
|
return envVar.ID
|
||||||
body := `[{"key":"NEW_KEY","value":"new_value"},{"key":"ANOTHER","value":"42"}]`
|
},
|
||||||
|
deletePath: func(appID string, resourceID int64) string {
|
||||||
|
return "/apps/" + appID + "/env/" + strconv.FormatInt(resourceID, 10) + "/delete"
|
||||||
|
},
|
||||||
|
chiParams: func(appID string, resourceID int64) map[string]string {
|
||||||
|
return map[string]string{"id": appID, "varID": strconv.FormatInt(resourceID, 10)}
|
||||||
|
},
|
||||||
|
handler: func(h *handlers.Handlers) http.HandlerFunc { return h.HandleEnvVarDelete() },
|
||||||
|
verifyFn: func(t *testing.T, tc *testContext, resourceID int64) {
|
||||||
|
t.Helper()
|
||||||
|
|
||||||
r := chi.NewRouter()
|
found, findErr := models.FindEnvVar(context.Background(), tc.database, resourceID)
|
||||||
r.Post("/apps/{id}/env", testCtx.handlers.HandleEnvVarSave())
|
require.NoError(t, findErr)
|
||||||
|
assert.NotNil(t, found, "env var should still exist after IDOR attempt")
|
||||||
request := httptest.NewRequest(
|
},
|
||||||
http.MethodPost,
|
})
|
||||||
"/apps/"+createdApp.ID+"/env",
|
|
||||||
strings.NewReader(body),
|
|
||||||
)
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
r.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
|
||||||
|
|
||||||
// Verify old env vars are gone and new ones exist
|
|
||||||
envVars, err := models.FindEnvVarsByAppID(
|
|
||||||
context.Background(), testCtx.database, createdApp.ID,
|
|
||||||
)
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.Len(t, envVars, 2)
|
|
||||||
|
|
||||||
keys := make(map[string]string)
|
|
||||||
for _, ev := range envVars {
|
|
||||||
keys[ev.Key] = ev.Value
|
|
||||||
}
|
|
||||||
|
|
||||||
assert.Equal(t, "new_value", keys["NEW_KEY"])
|
|
||||||
assert.Equal(t, "42", keys["ANOTHER"])
|
|
||||||
assert.Empty(t, keys["OLD_KEY"], "old env vars should be deleted")
|
|
||||||
assert.Empty(t, keys["REMOVE_ME"], "old env vars should be deleted")
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHandleEnvVarSaveAppNotFound tests that HandleEnvVarSave returns 404
|
|
||||||
// for a non-existent app.
|
|
||||||
func TestHandleEnvVarSaveAppNotFound(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
|
|
||||||
body := `[{"key":"KEY","value":"value"}]`
|
|
||||||
|
|
||||||
r := chi.NewRouter()
|
|
||||||
r.Post("/apps/{id}/env", testCtx.handlers.HandleEnvVarSave())
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/nonexistent-id/env",
|
|
||||||
strings.NewReader(body),
|
|
||||||
)
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
r.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusNotFound, recorder.Code)
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHandleEnvVarSaveEmptyKeyRejected verifies that submitting a JSON
|
|
||||||
// array containing an entry with an empty key returns 400.
|
|
||||||
func TestHandleEnvVarSaveEmptyKeyRejected(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
createdApp := createTestApp(t, testCtx, "envvar-emptykey-app")
|
|
||||||
|
|
||||||
body := `[{"key":"VALID_KEY","value":"ok"},{"key":"","value":"bad"}]`
|
|
||||||
|
|
||||||
r := chi.NewRouter()
|
|
||||||
r.Post("/apps/{id}/env", testCtx.handlers.HandleEnvVarSave())
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/"+createdApp.ID+"/env",
|
|
||||||
strings.NewReader(body),
|
|
||||||
)
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
r.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusBadRequest, recorder.Code)
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHandleEnvVarSaveDuplicateKeyRejected verifies that when the client
|
|
||||||
// sends duplicate keys, the server rejects them with 400 Bad Request.
|
|
||||||
func TestHandleEnvVarSaveDuplicateKeyRejected(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
createdApp := createTestApp(t, testCtx, "envvar-dedup-app")
|
|
||||||
|
|
||||||
// Send two entries with the same key — should be rejected
|
|
||||||
body := `[{"key":"FOO","value":"first"},{"key":"BAR","value":"bar"},{"key":"FOO","value":"second"}]`
|
|
||||||
|
|
||||||
r := chi.NewRouter()
|
|
||||||
r.Post("/apps/{id}/env", testCtx.handlers.HandleEnvVarSave())
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/"+createdApp.ID+"/env",
|
|
||||||
strings.NewReader(body),
|
|
||||||
)
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
r.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusBadRequest, recorder.Code)
|
|
||||||
assert.Contains(t, recorder.Body.String(), "duplicate environment variable key: FOO")
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHandleEnvVarSaveCrossAppIsolation verifies that posting env vars
|
|
||||||
// to appA's endpoint does not affect appB's env vars (IDOR prevention).
|
|
||||||
func TestHandleEnvVarSaveCrossAppIsolation(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
appA := createTestApp(t, testCtx, "envvar-iso-appA")
|
|
||||||
appB := createTestApp(t, testCtx, "envvar-iso-appB")
|
|
||||||
|
|
||||||
// Give appB some env vars
|
|
||||||
for _, kv := range [][2]string{{"B_KEY1", "b_val1"}, {"B_KEY2", "b_val2"}} {
|
|
||||||
ev := models.NewEnvVar(testCtx.database)
|
|
||||||
ev.AppID = appB.ID
|
|
||||||
ev.Key = kv[0]
|
|
||||||
ev.Value = kv[1]
|
|
||||||
require.NoError(t, ev.Save(context.Background()))
|
|
||||||
}
|
|
||||||
|
|
||||||
// POST new env vars to appA's endpoint
|
|
||||||
body := `[{"key":"A_KEY","value":"a_val"}]`
|
|
||||||
|
|
||||||
r := chi.NewRouter()
|
|
||||||
r.Post("/apps/{id}/env", testCtx.handlers.HandleEnvVarSave())
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/"+appA.ID+"/env",
|
|
||||||
strings.NewReader(body),
|
|
||||||
)
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
r.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
|
||||||
|
|
||||||
// Verify appA has exactly what we sent
|
|
||||||
appAVars, err := models.FindEnvVarsByAppID(
|
|
||||||
context.Background(), testCtx.database, appA.ID,
|
|
||||||
)
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.Len(t, appAVars, 1)
|
|
||||||
assert.Equal(t, "A_KEY", appAVars[0].Key)
|
|
||||||
|
|
||||||
// Verify appB's env vars are completely untouched
|
|
||||||
appBVars, err := models.FindEnvVarsByAppID(
|
|
||||||
context.Background(), testCtx.database, appB.ID,
|
|
||||||
)
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.Len(t, appBVars, 2, "appB env vars must not be affected")
|
|
||||||
|
|
||||||
bKeys := make(map[string]string)
|
|
||||||
for _, ev := range appBVars {
|
|
||||||
bKeys[ev.Key] = ev.Value
|
|
||||||
}
|
|
||||||
|
|
||||||
assert.Equal(t, "b_val1", bKeys["B_KEY1"])
|
|
||||||
assert.Equal(t, "b_val2", bKeys["B_KEY2"])
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHandleEnvVarSaveBodySizeLimit verifies that a request body
|
|
||||||
// exceeding the 1 MB limit is rejected.
|
|
||||||
func TestHandleEnvVarSaveBodySizeLimit(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
createdApp := createTestApp(t, testCtx, "envvar-sizelimit-app")
|
|
||||||
|
|
||||||
// Build a JSON body that exceeds 1 MB
|
|
||||||
// Each entry is ~30 bytes; 40000 entries ≈ 1.2 MB
|
|
||||||
var sb strings.Builder
|
|
||||||
|
|
||||||
sb.WriteString("[")
|
|
||||||
|
|
||||||
for i := range 40000 {
|
|
||||||
if i > 0 {
|
|
||||||
sb.WriteString(",")
|
|
||||||
}
|
|
||||||
|
|
||||||
sb.WriteString(`{"key":"K` + strconv.Itoa(i) + `","value":"val"}`)
|
|
||||||
}
|
|
||||||
|
|
||||||
sb.WriteString("]")
|
|
||||||
|
|
||||||
r := chi.NewRouter()
|
|
||||||
r.Post("/apps/{id}/env", testCtx.handlers.HandleEnvVarSave())
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/"+createdApp.ID+"/env",
|
|
||||||
strings.NewReader(sb.String()),
|
|
||||||
)
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
r.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusBadRequest, recorder.Code,
|
|
||||||
"oversized body should be rejected with 400")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestDeleteLabelOwnershipVerification tests that deleting a label
|
// TestDeleteLabelOwnershipVerification tests that deleting a label
|
||||||
// via another app's URL path returns 404 (IDOR prevention).
|
// via another app's URL path returns 404 (IDOR prevention).
|
||||||
func TestDeleteLabelOwnershipVerification(t *testing.T) {
|
func TestDeleteLabelOwnershipVerification(t *testing.T) { //nolint:dupl // intentionally similar IDOR test pattern
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
testOwnershipVerification(t, ownedResourceTestConfig{
|
testOwnershipVerification(t, ownedResourceTestConfig{
|
||||||
@@ -911,43 +695,41 @@ func TestDeletePortOwnershipVerification(t *testing.T) {
|
|||||||
assert.NotNil(t, found, "port should still exist after IDOR attempt")
|
assert.NotNil(t, found, "port should still exist after IDOR attempt")
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestHandleEnvVarSaveEmptyClears verifies that submitting an empty JSON
|
// TestHandleEnvVarDeleteUsesCorrectRouteParam verifies that HandleEnvVarDelete
|
||||||
// array deletes all existing env vars for the app.
|
// reads the "varID" chi URL parameter (matching the route definition {varID}),
|
||||||
func TestHandleEnvVarSaveEmptyClears(t *testing.T) {
|
// not a mismatched name like "envID".
|
||||||
|
func TestHandleEnvVarDeleteUsesCorrectRouteParam(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
testCtx := setupTestHandlers(t)
|
||||||
createdApp := createTestApp(t, testCtx, "envvar-clear-app")
|
|
||||||
|
|
||||||
// Create a pre-existing env var
|
createdApp := createTestApp(t, testCtx, "envdelete-param-app")
|
||||||
ev := models.NewEnvVar(testCtx.database)
|
|
||||||
ev.AppID = createdApp.ID
|
|
||||||
ev.Key = "DELETE_ME"
|
|
||||||
ev.Value = "gone"
|
|
||||||
require.NoError(t, ev.Save(context.Background()))
|
|
||||||
|
|
||||||
// Submit empty JSON array
|
envVar := models.NewEnvVar(testCtx.database)
|
||||||
|
envVar.AppID = createdApp.ID
|
||||||
|
envVar.Key = "DELETE_ME"
|
||||||
|
envVar.Value = "gone"
|
||||||
|
require.NoError(t, envVar.Save(context.Background()))
|
||||||
|
|
||||||
|
// Use chi router with the real route pattern to test param name
|
||||||
r := chi.NewRouter()
|
r := chi.NewRouter()
|
||||||
r.Post("/apps/{id}/env", testCtx.handlers.HandleEnvVarSave())
|
r.Post("/apps/{id}/env-vars/{varID}/delete", testCtx.handlers.HandleEnvVarDelete())
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
request := httptest.NewRequest(
|
||||||
http.MethodPost,
|
http.MethodPost,
|
||||||
"/apps/"+createdApp.ID+"/env",
|
"/apps/"+createdApp.ID+"/env-vars/"+strconv.FormatInt(envVar.ID, 10)+"/delete",
|
||||||
strings.NewReader("[]"),
|
nil,
|
||||||
)
|
)
|
||||||
request.Header.Set("Content-Type", "application/json")
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
recorder := httptest.NewRecorder()
|
||||||
|
|
||||||
r.ServeHTTP(recorder, request)
|
r.ServeHTTP(recorder, request)
|
||||||
|
|
||||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
assert.Equal(t, http.StatusSeeOther, recorder.Code)
|
||||||
|
|
||||||
// Verify all env vars are gone
|
// Verify the env var was actually deleted
|
||||||
envVars, err := models.FindEnvVarsByAppID(
|
found, findErr := models.FindEnvVar(context.Background(), testCtx.database, envVar.ID)
|
||||||
context.Background(), testCtx.database, createdApp.ID,
|
require.NoError(t, findErr)
|
||||||
)
|
assert.Nil(t, found, "env var should be deleted when using correct route param")
|
||||||
require.NoError(t, err)
|
|
||||||
assert.Empty(t, envVars, "all env vars should be deleted")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestHandleVolumeAddValidatesPaths verifies that HandleVolumeAdd validates
|
// TestHandleVolumeAddValidatesPaths verifies that HandleVolumeAdd validates
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package handlers_test
|
|||||||
import (
|
import (
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/handlers"
|
"git.eeqj.de/sneak/upaas/internal/handlers"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestValidateRepoURL(t *testing.T) {
|
func TestValidateRepoURL(t *testing.T) {
|
||||||
|
|||||||
@@ -1,195 +0,0 @@
|
|||||||
package handlers //nolint:testpackage // tests unexported parsing functions
|
|
||||||
|
|
||||||
import (
|
|
||||||
"database/sql"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
|
||||||
"github.com/stretchr/testify/require"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestParseOptionalFloat64(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
t.Run("empty string returns invalid", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalFloat64("")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.False(t, result.Valid)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("whitespace only returns invalid", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalFloat64(" ")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.False(t, result.Valid)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("valid float", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalFloat64("0.5")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.True(t, result.Valid)
|
|
||||||
assert.InDelta(t, 0.5, result.Float64, 0.001)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("valid integer", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalFloat64("2")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.True(t, result.Valid)
|
|
||||||
assert.InDelta(t, 2.0, result.Float64, 0.001)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("negative value rejected", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
_, err := parseOptionalFloat64("-1")
|
|
||||||
require.Error(t, err)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("zero value rejected", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
_, err := parseOptionalFloat64("0")
|
|
||||||
require.Error(t, err)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("non-numeric rejected", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
_, err := parseOptionalFloat64("abc")
|
|
||||||
require.Error(t, err)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestParseOptionalMemoryBytes(t *testing.T) { //nolint:funlen // table-driven test
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
t.Run("empty string returns invalid", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes("")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.False(t, result.Valid)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("whitespace only returns invalid", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes(" ")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.False(t, result.Valid)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("plain bytes", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes("536870912")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.True(t, result.Valid)
|
|
||||||
assert.Equal(t, int64(536870912), result.Int64)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("megabytes suffix", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes("256m")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.True(t, result.Valid)
|
|
||||||
assert.Equal(t, int64(256*1024*1024), result.Int64)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("megabytes suffix uppercase", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes("256M")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.True(t, result.Valid)
|
|
||||||
assert.Equal(t, int64(256*1024*1024), result.Int64)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("gigabytes suffix", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes("1g")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.True(t, result.Valid)
|
|
||||||
assert.Equal(t, int64(1024*1024*1024), result.Int64)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("kilobytes suffix", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes("512k")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.True(t, result.Valid)
|
|
||||||
assert.Equal(t, int64(512*1024), result.Int64)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("fractional gigabytes", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes("1.5g")
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.True(t, result.Valid)
|
|
||||||
assert.Equal(t, int64(1.5*1024*1024*1024), result.Int64)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("negative value rejected", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
_, err := parseOptionalMemoryBytes("-256m")
|
|
||||||
require.Error(t, err)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("zero value rejected", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
_, err := parseOptionalMemoryBytes("0")
|
|
||||||
require.Error(t, err)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("invalid string rejected", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
_, err := parseOptionalMemoryBytes("abc")
|
|
||||||
require.Error(t, err)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("negative plain bytes rejected", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
_, err := parseOptionalMemoryBytes("-100")
|
|
||||||
require.Error(t, err)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestAppResourceLimitsRoundTrip(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
// Test that parsing and formatting are consistent
|
|
||||||
tests := []struct {
|
|
||||||
input string
|
|
||||||
expected sql.NullInt64
|
|
||||||
format string
|
|
||||||
}{
|
|
||||||
{"256m", sql.NullInt64{Int64: 256 * 1024 * 1024, Valid: true}, "256m"},
|
|
||||||
{"1g", sql.NullInt64{Int64: 1024 * 1024 * 1024, Valid: true}, "1g"},
|
|
||||||
{"512k", sql.NullInt64{Int64: 512 * 1024, Valid: true}, "512k"},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.input, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
result, err := parseOptionalMemoryBytes(tt.input)
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.Equal(t, tt.expected, result)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -3,7 +3,7 @@ package handlers_test
|
|||||||
import (
|
import (
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/handlers"
|
"git.eeqj.de/sneak/upaas/internal/handlers"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestSanitizeLogs(t *testing.T) { //nolint:funlen // table-driven tests
|
func TestSanitizeLogs(t *testing.T) { //nolint:funlen // table-driven tests
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package handlers
|
|||||||
import (
|
import (
|
||||||
"net/http"
|
"net/http"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/templates"
|
"git.eeqj.de/sneak/upaas/templates"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package handlers_test
|
|||||||
import (
|
import (
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/handlers"
|
"git.eeqj.de/sneak/upaas/internal/handlers"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestSanitizeTail(t *testing.T) {
|
func TestSanitizeTail(t *testing.T) {
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import (
|
|||||||
|
|
||||||
"github.com/go-chi/chi/v5"
|
"github.com/go-chi/chi/v5"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
)
|
)
|
||||||
|
|
||||||
// maxWebhookBodySize is the maximum allowed size of a webhook request body (1MB).
|
// maxWebhookBodySize is the maximum allowed size of a webhook request body (1MB).
|
||||||
|
|||||||
@@ -1,56 +0,0 @@
|
|||||||
package handlers
|
|
||||||
|
|
||||||
import (
|
|
||||||
"net/http"
|
|
||||||
|
|
||||||
"github.com/go-chi/chi/v5"
|
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
|
||||||
"sneak.berlin/go/upaas/templates"
|
|
||||||
)
|
|
||||||
|
|
||||||
// webhookEventsLimit is the number of webhook events to show in history.
|
|
||||||
const webhookEventsLimit = 100
|
|
||||||
|
|
||||||
// HandleAppWebhookEvents returns the webhook event history handler.
|
|
||||||
func (h *Handlers) HandleAppWebhookEvents() http.HandlerFunc {
|
|
||||||
tmpl := templates.GetParsed()
|
|
||||||
|
|
||||||
return func(writer http.ResponseWriter, request *http.Request) {
|
|
||||||
appID := chi.URLParam(request, "id")
|
|
||||||
|
|
||||||
application, findErr := models.FindApp(request.Context(), h.db, appID)
|
|
||||||
if findErr != nil {
|
|
||||||
h.log.Error("failed to find app", "error", findErr)
|
|
||||||
http.Error(writer, "Internal Server Error", http.StatusInternalServerError)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if application == nil {
|
|
||||||
http.NotFound(writer, request)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
events, eventsErr := application.GetWebhookEvents(
|
|
||||||
request.Context(),
|
|
||||||
webhookEventsLimit,
|
|
||||||
)
|
|
||||||
if eventsErr != nil {
|
|
||||||
h.log.Error("failed to get webhook events",
|
|
||||||
"error", eventsErr,
|
|
||||||
"app", appID,
|
|
||||||
)
|
|
||||||
|
|
||||||
events = []*models.WebhookEvent{}
|
|
||||||
}
|
|
||||||
|
|
||||||
data := h.addGlobals(map[string]any{
|
|
||||||
"App": application,
|
|
||||||
"Events": events,
|
|
||||||
}, request)
|
|
||||||
|
|
||||||
h.renderTemplate(writer, tmpl, "webhook_events.html", data)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -8,10 +8,10 @@ import (
|
|||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Params contains dependencies for Healthcheck.
|
// Params contains dependencies for Healthcheck.
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ import (
|
|||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Params contains dependencies for Logger.
|
// Params contains dependencies for Logger.
|
||||||
|
|||||||
@@ -1,11 +0,0 @@
|
|||||||
package logger
|
|
||||||
|
|
||||||
import "log/slog"
|
|
||||||
|
|
||||||
// NewForTest creates a Logger wrapping the given slog.Logger, for use in tests.
|
|
||||||
func NewForTest(log *slog.Logger) *Logger {
|
|
||||||
return &Logger{
|
|
||||||
log: log,
|
|
||||||
level: new(slog.LevelVar),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
)
|
)
|
||||||
|
|
||||||
//nolint:gosec // test credentials
|
//nolint:gosec // test credentials
|
||||||
|
|||||||
@@ -18,10 +18,10 @@ import (
|
|||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
"golang.org/x/time/rate"
|
"golang.org/x/time/rate"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/service/auth"
|
"git.eeqj.de/sneak/upaas/internal/service/auth"
|
||||||
)
|
)
|
||||||
|
|
||||||
// corsMaxAge is the maximum age for CORS preflight responses in seconds.
|
// corsMaxAge is the maximum age for CORS preflight responses in seconds.
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ import (
|
|||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
)
|
)
|
||||||
|
|
||||||
func newTestMiddleware(t *testing.T) *Middleware {
|
func newTestMiddleware(t *testing.T) *Middleware {
|
||||||
|
|||||||
@@ -7,15 +7,14 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
// appColumns is the standard column list for app queries.
|
// appColumns is the standard column list for app queries.
|
||||||
const appColumns = `id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
const appColumns = `id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
||||||
ssh_private_key, ssh_public_key, image_id, status,
|
ssh_private_key, ssh_public_key, image_id, status,
|
||||||
docker_network, ntfy_topic, slack_webhook, webhook_secret_hash,
|
docker_network, ntfy_topic, slack_webhook, webhook_secret_hash,
|
||||||
previous_image_id, cpu_limit, memory_limit,
|
previous_image_id, created_at, updated_at`
|
||||||
created_at, updated_at`
|
|
||||||
|
|
||||||
// AppStatus represents the status of an app.
|
// AppStatus represents the status of an app.
|
||||||
type AppStatus string
|
type AppStatus string
|
||||||
@@ -48,8 +47,6 @@ type App struct {
|
|||||||
DockerNetwork sql.NullString
|
DockerNetwork sql.NullString
|
||||||
NtfyTopic sql.NullString
|
NtfyTopic sql.NullString
|
||||||
SlackWebhook sql.NullString
|
SlackWebhook sql.NullString
|
||||||
CPULimit sql.NullFloat64
|
|
||||||
MemoryLimit sql.NullInt64
|
|
||||||
CreatedAt time.Time
|
CreatedAt time.Time
|
||||||
UpdatedAt time.Time
|
UpdatedAt time.Time
|
||||||
}
|
}
|
||||||
@@ -145,14 +142,14 @@ func (a *App) insert(ctx context.Context) error {
|
|||||||
id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
||||||
ssh_private_key, ssh_public_key, image_id, status,
|
ssh_private_key, ssh_public_key, image_id, status,
|
||||||
docker_network, ntfy_topic, slack_webhook, webhook_secret_hash,
|
docker_network, ntfy_topic, slack_webhook, webhook_secret_hash,
|
||||||
previous_image_id, cpu_limit, memory_limit
|
previous_image_id
|
||||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)`
|
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)`
|
||||||
|
|
||||||
_, err := a.db.Exec(ctx, query,
|
_, err := a.db.Exec(ctx, query,
|
||||||
a.ID, a.Name, a.RepoURL, a.Branch, a.DockerfilePath, a.WebhookSecret,
|
a.ID, a.Name, a.RepoURL, a.Branch, a.DockerfilePath, a.WebhookSecret,
|
||||||
a.SSHPrivateKey, a.SSHPublicKey, a.ImageID, a.Status,
|
a.SSHPrivateKey, a.SSHPublicKey, a.ImageID, a.Status,
|
||||||
a.DockerNetwork, a.NtfyTopic, a.SlackWebhook, a.WebhookSecretHash,
|
a.DockerNetwork, a.NtfyTopic, a.SlackWebhook, a.WebhookSecretHash,
|
||||||
a.PreviousImageID, a.CPULimit, a.MemoryLimit,
|
a.PreviousImageID,
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@@ -168,7 +165,6 @@ func (a *App) update(ctx context.Context) error {
|
|||||||
image_id = ?, status = ?,
|
image_id = ?, status = ?,
|
||||||
docker_network = ?, ntfy_topic = ?, slack_webhook = ?,
|
docker_network = ?, ntfy_topic = ?, slack_webhook = ?,
|
||||||
previous_image_id = ?,
|
previous_image_id = ?,
|
||||||
cpu_limit = ?, memory_limit = ?,
|
|
||||||
updated_at = CURRENT_TIMESTAMP
|
updated_at = CURRENT_TIMESTAMP
|
||||||
WHERE id = ?`
|
WHERE id = ?`
|
||||||
|
|
||||||
@@ -177,7 +173,6 @@ func (a *App) update(ctx context.Context) error {
|
|||||||
a.ImageID, a.Status,
|
a.ImageID, a.Status,
|
||||||
a.DockerNetwork, a.NtfyTopic, a.SlackWebhook,
|
a.DockerNetwork, a.NtfyTopic, a.SlackWebhook,
|
||||||
a.PreviousImageID,
|
a.PreviousImageID,
|
||||||
a.CPULimit, a.MemoryLimit,
|
|
||||||
a.ID,
|
a.ID,
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -193,7 +188,6 @@ func (a *App) scan(row *sql.Row) error {
|
|||||||
&a.DockerNetwork, &a.NtfyTopic, &a.SlackWebhook,
|
&a.DockerNetwork, &a.NtfyTopic, &a.SlackWebhook,
|
||||||
&a.WebhookSecretHash,
|
&a.WebhookSecretHash,
|
||||||
&a.PreviousImageID,
|
&a.PreviousImageID,
|
||||||
&a.CPULimit, &a.MemoryLimit,
|
|
||||||
&a.CreatedAt, &a.UpdatedAt,
|
&a.CreatedAt, &a.UpdatedAt,
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
@@ -212,7 +206,6 @@ func scanApps(appDB *database.Database, rows *sql.Rows) ([]*App, error) {
|
|||||||
&app.DockerNetwork, &app.NtfyTopic, &app.SlackWebhook,
|
&app.DockerNetwork, &app.NtfyTopic, &app.SlackWebhook,
|
||||||
&app.WebhookSecretHash,
|
&app.WebhookSecretHash,
|
||||||
&app.PreviousImageID,
|
&app.PreviousImageID,
|
||||||
&app.CPULimit, &app.MemoryLimit,
|
|
||||||
&app.CreatedAt, &app.UpdatedAt,
|
&app.CreatedAt, &app.UpdatedAt,
|
||||||
)
|
)
|
||||||
if scanErr != nil {
|
if scanErr != nil {
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
// DeploymentStatus represents the status of a deployment.
|
// DeploymentStatus represents the status of a deployment.
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
//nolint:dupl // Active Record pattern - similar structure to label.go is intentional
|
||||||
package models
|
package models
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@@ -6,7 +7,7 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
// EnvVar represents an environment variable for an app.
|
// EnvVar represents an environment variable for an app.
|
||||||
@@ -128,48 +129,13 @@ func FindEnvVarsByAppID(
|
|||||||
return envVars, rows.Err()
|
return envVars, rows.Err()
|
||||||
}
|
}
|
||||||
|
|
||||||
// EnvVarPair is a key-value pair for bulk env var operations.
|
// DeleteEnvVarsByAppID deletes all env vars for an app.
|
||||||
type EnvVarPair struct {
|
func DeleteEnvVarsByAppID(
|
||||||
Key string
|
|
||||||
Value string
|
|
||||||
}
|
|
||||||
|
|
||||||
// ReplaceEnvVarsByAppID atomically replaces all env vars for an app
|
|
||||||
// within a single database transaction. It deletes all existing env
|
|
||||||
// vars and inserts the provided pairs. If any operation fails, the
|
|
||||||
// entire transaction is rolled back.
|
|
||||||
func ReplaceEnvVarsByAppID(
|
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
db *database.Database,
|
db *database.Database,
|
||||||
appID string,
|
appID string,
|
||||||
pairs []EnvVarPair,
|
|
||||||
) error {
|
) error {
|
||||||
tx, err := db.BeginTx(ctx, nil)
|
_, err := db.Exec(ctx, "DELETE FROM app_env_vars WHERE app_id = ?", appID)
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("beginning transaction: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
defer func() { _ = tx.Rollback() }()
|
return err
|
||||||
|
|
||||||
_, err = tx.ExecContext(ctx, "DELETE FROM app_env_vars WHERE app_id = ?", appID)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("deleting env vars: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, p := range pairs {
|
|
||||||
_, err = tx.ExecContext(ctx,
|
|
||||||
"INSERT INTO app_env_vars (app_id, key, value) VALUES (?, ?, ?)",
|
|
||||||
appID, p.Key, p.Value,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("inserting env var %q: %w", p.Key, err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
err = tx.Commit()
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("committing transaction: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
//nolint:dupl // Active Record pattern - similar structure to env_var.go is intentional
|
||||||
package models
|
package models
|
||||||
|
|
||||||
import (
|
import (
|
||||||
@@ -6,7 +7,7 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Label represents a Docker label for an app container.
|
// Label represents a Docker label for an app container.
|
||||||
|
|||||||
@@ -10,11 +10,11 @@ import (
|
|||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Test constants to satisfy goconst linter.
|
// Test constants to satisfy goconst linter.
|
||||||
@@ -781,96 +781,6 @@ func TestCascadeDelete(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
// Resource Limits Tests.
|
|
||||||
|
|
||||||
func TestAppResourceLimits(t *testing.T) { //nolint:funlen // integration test with multiple subtests
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
t.Run("saves and loads CPU limit", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testDB, cleanup := setupTestDB(t)
|
|
||||||
defer cleanup()
|
|
||||||
|
|
||||||
app := createTestApp(t, testDB)
|
|
||||||
|
|
||||||
app.CPULimit = sql.NullFloat64{Float64: 0.5, Valid: true}
|
|
||||||
|
|
||||||
err := app.Save(context.Background())
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
found, err := models.FindApp(context.Background(), testDB, app.ID)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.NotNil(t, found)
|
|
||||||
assert.True(t, found.CPULimit.Valid)
|
|
||||||
assert.InDelta(t, 0.5, found.CPULimit.Float64, 0.001)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("saves and loads memory limit", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testDB, cleanup := setupTestDB(t)
|
|
||||||
defer cleanup()
|
|
||||||
|
|
||||||
app := createTestApp(t, testDB)
|
|
||||||
|
|
||||||
app.MemoryLimit = sql.NullInt64{Int64: 536870912, Valid: true} // 512m
|
|
||||||
|
|
||||||
err := app.Save(context.Background())
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
found, err := models.FindApp(context.Background(), testDB, app.ID)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.NotNil(t, found)
|
|
||||||
assert.True(t, found.MemoryLimit.Valid)
|
|
||||||
assert.Equal(t, int64(536870912), found.MemoryLimit.Int64)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("null limits by default", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testDB, cleanup := setupTestDB(t)
|
|
||||||
defer cleanup()
|
|
||||||
|
|
||||||
app := createTestApp(t, testDB)
|
|
||||||
|
|
||||||
found, err := models.FindApp(context.Background(), testDB, app.ID)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.NotNil(t, found)
|
|
||||||
assert.False(t, found.CPULimit.Valid)
|
|
||||||
assert.False(t, found.MemoryLimit.Valid)
|
|
||||||
})
|
|
||||||
|
|
||||||
t.Run("clears limits when set to null", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testDB, cleanup := setupTestDB(t)
|
|
||||||
defer cleanup()
|
|
||||||
|
|
||||||
app := createTestApp(t, testDB)
|
|
||||||
|
|
||||||
// Set limits
|
|
||||||
app.CPULimit = sql.NullFloat64{Float64: 1.0, Valid: true}
|
|
||||||
app.MemoryLimit = sql.NullInt64{Int64: 1073741824, Valid: true} // 1g
|
|
||||||
|
|
||||||
err := app.Save(context.Background())
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
// Clear limits
|
|
||||||
app.CPULimit = sql.NullFloat64{}
|
|
||||||
app.MemoryLimit = sql.NullInt64{}
|
|
||||||
|
|
||||||
err = app.Save(context.Background())
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
found, err := models.FindApp(context.Background(), testDB, app.ID)
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.NotNil(t, found)
|
|
||||||
assert.False(t, found.CPULimit.Valid)
|
|
||||||
assert.False(t, found.MemoryLimit.Valid)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
// Helper function to create a test app.
|
// Helper function to create a test app.
|
||||||
func createTestApp(t *testing.T, testDB *database.Database) *models.App {
|
func createTestApp(t *testing.T, testDB *database.Database) *models.App {
|
||||||
t.Helper()
|
t.Helper()
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
// PortProtocol represents the protocol for a port mapping.
|
// PortProtocol represents the protocol for a port mapping.
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
// User represents a user in the system.
|
// User represents a user in the system.
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Volume represents a volume mount for an app container.
|
// Volume represents a volume mount for an app container.
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
)
|
)
|
||||||
|
|
||||||
// WebhookEvent represents a received webhook event.
|
// WebhookEvent represents a received webhook event.
|
||||||
@@ -52,20 +52,6 @@ func (w *WebhookEvent) Reload(ctx context.Context) error {
|
|||||||
return w.scan(row)
|
return w.scan(row)
|
||||||
}
|
}
|
||||||
|
|
||||||
// ShortCommit returns a truncated commit SHA for display.
|
|
||||||
func (w *WebhookEvent) ShortCommit() string {
|
|
||||||
if !w.CommitSHA.Valid {
|
|
||||||
return ""
|
|
||||||
}
|
|
||||||
|
|
||||||
sha := w.CommitSHA.String
|
|
||||||
if len(sha) > shortCommitLength {
|
|
||||||
return sha[:shortCommitLength]
|
|
||||||
}
|
|
||||||
|
|
||||||
return sha
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *WebhookEvent) insert(ctx context.Context) error {
|
func (w *WebhookEvent) insert(ctx context.Context) error {
|
||||||
query := `
|
query := `
|
||||||
INSERT INTO webhook_events (
|
INSERT INTO webhook_events (
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ import (
|
|||||||
chimw "github.com/go-chi/chi/v5/middleware"
|
chimw "github.com/go-chi/chi/v5/middleware"
|
||||||
"github.com/prometheus/client_golang/prometheus/promhttp"
|
"github.com/prometheus/client_golang/prometheus/promhttp"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/static"
|
"git.eeqj.de/sneak/upaas/static"
|
||||||
)
|
)
|
||||||
|
|
||||||
// requestTimeout is the maximum duration for handling a request.
|
// requestTimeout is the maximum duration for handling a request.
|
||||||
@@ -70,7 +70,6 @@ func (s *Server) SetupRoutes() {
|
|||||||
r.Post("/apps/{id}/deploy", s.handlers.HandleAppDeploy())
|
r.Post("/apps/{id}/deploy", s.handlers.HandleAppDeploy())
|
||||||
r.Post("/apps/{id}/deployments/cancel", s.handlers.HandleCancelDeploy())
|
r.Post("/apps/{id}/deployments/cancel", s.handlers.HandleCancelDeploy())
|
||||||
r.Get("/apps/{id}/deployments", s.handlers.HandleAppDeployments())
|
r.Get("/apps/{id}/deployments", s.handlers.HandleAppDeployments())
|
||||||
r.Get("/apps/{id}/webhooks", s.handlers.HandleAppWebhookEvents())
|
|
||||||
r.Get("/apps/{id}/deployments/{deploymentID}/logs", s.handlers.HandleDeploymentLogsAPI())
|
r.Get("/apps/{id}/deployments/{deploymentID}/logs", s.handlers.HandleDeploymentLogsAPI())
|
||||||
r.Get("/apps/{id}/deployments/{deploymentID}/download", s.handlers.HandleDeploymentLogDownload())
|
r.Get("/apps/{id}/deployments/{deploymentID}/download", s.handlers.HandleDeploymentLogDownload())
|
||||||
r.Get("/apps/{id}/logs", s.handlers.HandleAppLogs())
|
r.Get("/apps/{id}/logs", s.handlers.HandleAppLogs())
|
||||||
@@ -82,8 +81,10 @@ func (s *Server) SetupRoutes() {
|
|||||||
r.Post("/apps/{id}/stop", s.handlers.HandleAppStop())
|
r.Post("/apps/{id}/stop", s.handlers.HandleAppStop())
|
||||||
r.Post("/apps/{id}/start", s.handlers.HandleAppStart())
|
r.Post("/apps/{id}/start", s.handlers.HandleAppStart())
|
||||||
|
|
||||||
// Environment variables (monolithic bulk save)
|
// Environment variables
|
||||||
r.Post("/apps/{id}/env", s.handlers.HandleEnvVarSave())
|
r.Post("/apps/{id}/env-vars", s.handlers.HandleEnvVarAdd())
|
||||||
|
r.Post("/apps/{id}/env-vars/{varID}/edit", s.handlers.HandleEnvVarEdit())
|
||||||
|
r.Post("/apps/{id}/env-vars/{varID}/delete", s.handlers.HandleEnvVarDelete())
|
||||||
|
|
||||||
// Labels
|
// Labels
|
||||||
r.Post("/apps/{id}/labels", s.handlers.HandleLabelAdd())
|
r.Post("/apps/{id}/labels", s.handlers.HandleLabelAdd())
|
||||||
|
|||||||
@@ -12,11 +12,11 @@ import (
|
|||||||
"github.com/go-chi/chi/v5"
|
"github.com/go-chi/chi/v5"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/handlers"
|
"git.eeqj.de/sneak/upaas/internal/handlers"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/middleware"
|
"git.eeqj.de/sneak/upaas/internal/middleware"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Params contains dependencies for Server.
|
// Params contains dependencies for Server.
|
||||||
|
|||||||
@@ -14,10 +14,10 @@ import (
|
|||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
"sneak.berlin/go/upaas/internal/ssh"
|
"git.eeqj.de/sneak/upaas/internal/ssh"
|
||||||
)
|
)
|
||||||
|
|
||||||
// ServiceParams contains dependencies for Service.
|
// ServiceParams contains dependencies for Service.
|
||||||
|
|||||||
@@ -8,12 +8,12 @@ import (
|
|||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
"sneak.berlin/go/upaas/internal/service/app"
|
"git.eeqj.de/sneak/upaas/internal/service/app"
|
||||||
)
|
)
|
||||||
|
|
||||||
func setupTestService(t *testing.T) (*app.Service, func()) {
|
func setupTestService(t *testing.T) (*app.Service, func()) {
|
||||||
|
|||||||
@@ -15,10 +15,10 @@ import (
|
|||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
"golang.org/x/crypto/argon2"
|
"golang.org/x/crypto/argon2"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -12,11 +12,11 @@ import (
|
|||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/service/auth"
|
"git.eeqj.de/sneak/upaas/internal/service/auth"
|
||||||
)
|
)
|
||||||
|
|
||||||
func setupTestService(t *testing.T) (*auth.Service, func()) {
|
func setupTestService(t *testing.T) (*auth.Service, func()) {
|
||||||
|
|||||||
@@ -17,12 +17,12 @@ import (
|
|||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/docker"
|
"git.eeqj.de/sneak/upaas/internal/docker"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
"sneak.berlin/go/upaas/internal/service/notify"
|
"git.eeqj.de/sneak/upaas/internal/service/notify"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Time constants.
|
// Time constants.
|
||||||
@@ -417,13 +417,15 @@ func (svc *Service) executeRollback(
|
|||||||
|
|
||||||
svc.removeOldContainer(ctx, app, deployment)
|
svc.removeOldContainer(ctx, app, deployment)
|
||||||
|
|
||||||
rollbackOpts, err := svc.buildContainerOptions(ctx, app, docker.ImageID(previousImageID))
|
rollbackOpts, err := svc.buildContainerOptions(ctx, app, deployment.ID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
svc.failDeployment(bgCtx, app, deployment, err)
|
svc.failDeployment(bgCtx, app, deployment, err)
|
||||||
|
|
||||||
return fmt.Errorf("failed to build container options: %w", err)
|
return fmt.Errorf("failed to build container options: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
rollbackOpts.Image = previousImageID
|
||||||
|
|
||||||
containerID, err := svc.docker.CreateContainer(ctx, rollbackOpts)
|
containerID, err := svc.docker.CreateContainer(ctx, rollbackOpts)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
svc.failDeployment(bgCtx, app, deployment, fmt.Errorf("failed to create rollback container: %w", err))
|
svc.failDeployment(bgCtx, app, deployment, fmt.Errorf("failed to create rollback container: %w", err))
|
||||||
@@ -431,8 +433,8 @@ func (svc *Service) executeRollback(
|
|||||||
return fmt.Errorf("failed to create rollback container: %w", err)
|
return fmt.Errorf("failed to create rollback container: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
deployment.ContainerID = sql.NullString{String: containerID.String(), Valid: true}
|
deployment.ContainerID = sql.NullString{String: containerID, Valid: true}
|
||||||
_ = deployment.AppendLog(bgCtx, "Rollback container created: "+containerID.String())
|
_ = deployment.AppendLog(bgCtx, "Rollback container created: "+containerID)
|
||||||
|
|
||||||
startErr := svc.docker.StartContainer(ctx, containerID)
|
startErr := svc.docker.StartContainer(ctx, containerID)
|
||||||
if startErr != nil {
|
if startErr != nil {
|
||||||
@@ -482,7 +484,7 @@ func (svc *Service) runBuildAndDeploy(
|
|||||||
svc.notify.NotifyBuildSuccess(bgCtx, app, deployment)
|
svc.notify.NotifyBuildSuccess(bgCtx, app, deployment)
|
||||||
|
|
||||||
// Deploy phase with timeout
|
// Deploy phase with timeout
|
||||||
err = svc.deployContainerWithTimeout(deployCtx, app, deployment, imageID)
|
err = svc.deployContainerWithTimeout(deployCtx, app, deployment)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment, imageID)
|
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment, imageID)
|
||||||
if cancelErr != nil {
|
if cancelErr != nil {
|
||||||
@@ -514,7 +516,7 @@ func (svc *Service) buildImageWithTimeout(
|
|||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
) (docker.ImageID, error) {
|
) (string, error) {
|
||||||
buildCtx, cancel := context.WithTimeout(ctx, buildTimeout)
|
buildCtx, cancel := context.WithTimeout(ctx, buildTimeout)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
|
|
||||||
@@ -539,7 +541,6 @@ func (svc *Service) deployContainerWithTimeout(
|
|||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
imageID docker.ImageID,
|
|
||||||
) error {
|
) error {
|
||||||
deployCtx, cancel := context.WithTimeout(ctx, deployTimeout)
|
deployCtx, cancel := context.WithTimeout(ctx, deployTimeout)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
@@ -553,7 +554,7 @@ func (svc *Service) deployContainerWithTimeout(
|
|||||||
svc.removeOldContainer(deployCtx, app, deployment)
|
svc.removeOldContainer(deployCtx, app, deployment)
|
||||||
|
|
||||||
// Create and start the new container
|
// Create and start the new container
|
||||||
_, err = svc.createAndStartContainer(deployCtx, app, deployment, imageID)
|
_, err = svc.createAndStartContainer(deployCtx, app, deployment)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if errors.Is(deployCtx.Err(), context.DeadlineExceeded) {
|
if errors.Is(deployCtx.Err(), context.DeadlineExceeded) {
|
||||||
timeoutErr := fmt.Errorf("%w after %v", ErrDeployTimeout, deployTimeout)
|
timeoutErr := fmt.Errorf("%w after %v", ErrDeployTimeout, deployTimeout)
|
||||||
@@ -667,7 +668,7 @@ func (svc *Service) checkCancelled(
|
|||||||
bgCtx context.Context,
|
bgCtx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
imageID docker.ImageID,
|
imageID string,
|
||||||
) error {
|
) error {
|
||||||
if !errors.Is(deployCtx.Err(), context.Canceled) {
|
if !errors.Is(deployCtx.Err(), context.Canceled) {
|
||||||
return nil
|
return nil
|
||||||
@@ -687,7 +688,7 @@ func (svc *Service) cleanupCancelledDeploy(
|
|||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
imageID docker.ImageID,
|
imageID string,
|
||||||
) {
|
) {
|
||||||
// Clean up the intermediate Docker image if one was built
|
// Clean up the intermediate Docker image if one was built
|
||||||
if imageID != "" {
|
if imageID != "" {
|
||||||
@@ -695,11 +696,11 @@ func (svc *Service) cleanupCancelledDeploy(
|
|||||||
if removeErr != nil {
|
if removeErr != nil {
|
||||||
svc.log.Error("failed to remove image from cancelled deploy",
|
svc.log.Error("failed to remove image from cancelled deploy",
|
||||||
"error", removeErr, "app", app.Name, "image", imageID)
|
"error", removeErr, "app", app.Name, "image", imageID)
|
||||||
_ = deployment.AppendLog(ctx, "WARNING: failed to clean up image "+imageID.String()+": "+removeErr.Error())
|
_ = deployment.AppendLog(ctx, "WARNING: failed to clean up image "+imageID+": "+removeErr.Error())
|
||||||
} else {
|
} else {
|
||||||
svc.log.Info("cleaned up image from cancelled deploy",
|
svc.log.Info("cleaned up image from cancelled deploy",
|
||||||
"app", app.Name, "image", imageID)
|
"app", app.Name, "image", imageID)
|
||||||
_ = deployment.AppendLog(ctx, "Cleaned up intermediate image: "+imageID.String())
|
_ = deployment.AppendLog(ctx, "Cleaned up intermediate image: "+imageID)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -816,7 +817,7 @@ func (svc *Service) buildImage(
|
|||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
) (docker.ImageID, error) {
|
) (string, error) {
|
||||||
workDir, cleanup, err := svc.cloneRepository(ctx, app, deployment)
|
workDir, cleanup, err := svc.cloneRepository(ctx, app, deployment)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
@@ -850,8 +851,8 @@ func (svc *Service) buildImage(
|
|||||||
return "", fmt.Errorf("failed to build image: %w", err)
|
return "", fmt.Errorf("failed to build image: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
deployment.ImageID = sql.NullString{String: imageID.String(), Valid: true}
|
deployment.ImageID = sql.NullString{String: imageID, Valid: true}
|
||||||
_ = deployment.AppendLog(ctx, "Image built: "+imageID.String())
|
_ = deployment.AppendLog(ctx, "Image built: "+imageID)
|
||||||
|
|
||||||
return imageID, nil
|
return imageID, nil
|
||||||
}
|
}
|
||||||
@@ -1009,16 +1010,15 @@ func (svc *Service) removeOldContainer(
|
|||||||
svc.log.Warn("failed to remove old container", "error", removeErr)
|
svc.log.Warn("failed to remove old container", "error", removeErr)
|
||||||
}
|
}
|
||||||
|
|
||||||
_ = deployment.AppendLog(ctx, "Old container removed: "+string(containerInfo.ID[:12]))
|
_ = deployment.AppendLog(ctx, "Old container removed: "+containerInfo.ID[:12])
|
||||||
}
|
}
|
||||||
|
|
||||||
func (svc *Service) createAndStartContainer(
|
func (svc *Service) createAndStartContainer(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
imageID docker.ImageID,
|
) (string, error) {
|
||||||
) (docker.ContainerID, error) {
|
containerOpts, err := svc.buildContainerOptions(ctx, app, deployment.ID)
|
||||||
containerOpts, err := svc.buildContainerOptions(ctx, app, imageID)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
svc.failDeployment(ctx, app, deployment, err)
|
svc.failDeployment(ctx, app, deployment, err)
|
||||||
|
|
||||||
@@ -1038,8 +1038,8 @@ func (svc *Service) createAndStartContainer(
|
|||||||
return "", fmt.Errorf("failed to create container: %w", err)
|
return "", fmt.Errorf("failed to create container: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
deployment.ContainerID = sql.NullString{String: containerID.String(), Valid: true}
|
deployment.ContainerID = sql.NullString{String: containerID, Valid: true}
|
||||||
_ = deployment.AppendLog(ctx, "Container created: "+containerID.String())
|
_ = deployment.AppendLog(ctx, "Container created: "+containerID)
|
||||||
|
|
||||||
startErr := svc.docker.StartContainer(ctx, containerID)
|
startErr := svc.docker.StartContainer(ctx, containerID)
|
||||||
if startErr != nil {
|
if startErr != nil {
|
||||||
@@ -1062,7 +1062,7 @@ func (svc *Service) createAndStartContainer(
|
|||||||
func (svc *Service) buildContainerOptions(
|
func (svc *Service) buildContainerOptions(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
imageID docker.ImageID,
|
deploymentID int64,
|
||||||
) (docker.CreateContainerOptions, error) {
|
) (docker.CreateContainerOptions, error) {
|
||||||
envVars, err := app.GetEnvVars(ctx)
|
envVars, err := app.GetEnvVars(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -1094,28 +1094,14 @@ func (svc *Service) buildContainerOptions(
|
|||||||
network = app.DockerNetwork.String
|
network = app.DockerNetwork.String
|
||||||
}
|
}
|
||||||
|
|
||||||
var cpuLimit float64
|
|
||||||
|
|
||||||
if app.CPULimit.Valid {
|
|
||||||
cpuLimit = app.CPULimit.Float64
|
|
||||||
}
|
|
||||||
|
|
||||||
var memoryLimit int64
|
|
||||||
|
|
||||||
if app.MemoryLimit.Valid {
|
|
||||||
memoryLimit = app.MemoryLimit.Int64
|
|
||||||
}
|
|
||||||
|
|
||||||
return docker.CreateContainerOptions{
|
return docker.CreateContainerOptions{
|
||||||
Name: "upaas-" + app.Name,
|
Name: "upaas-" + app.Name,
|
||||||
Image: imageID.String(),
|
Image: fmt.Sprintf("upaas-%s:%d", app.Name, deploymentID),
|
||||||
Env: envMap,
|
Env: envMap,
|
||||||
Labels: buildLabelMap(app, labels),
|
Labels: buildLabelMap(app, labels),
|
||||||
Volumes: buildVolumeMounts(volumes),
|
Volumes: buildVolumeMounts(volumes),
|
||||||
Ports: buildPortMappings(ports),
|
Ports: buildPortMappings(ports),
|
||||||
Network: network,
|
Network: network,
|
||||||
CPULimit: cpuLimit,
|
|
||||||
MemoryLimit: memoryLimit,
|
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1160,9 +1146,9 @@ func buildPortMappings(ports []*models.Port) []docker.PortMapping {
|
|||||||
func (svc *Service) updateAppRunning(
|
func (svc *Service) updateAppRunning(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
imageID docker.ImageID,
|
imageID string,
|
||||||
) error {
|
) error {
|
||||||
app.ImageID = sql.NullString{String: imageID.String(), Valid: true}
|
app.ImageID = sql.NullString{String: imageID, Valid: true}
|
||||||
app.Status = models.AppStatusRunning
|
app.Status = models.AppStatusRunning
|
||||||
|
|
||||||
saveErr := app.Save(ctx)
|
saveErr := app.Save(ctx)
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ import (
|
|||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/service/deploy"
|
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestCancelActiveDeploy_NoExisting(t *testing.T) {
|
func TestCancelActiveDeploy_NoExisting(t *testing.T) {
|
||||||
|
|||||||
@@ -10,8 +10,8 @@ import (
|
|||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/service/deploy"
|
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestCleanupCancelledDeploy_RemovesBuildDir(t *testing.T) {
|
func TestCleanupCancelledDeploy_RemovesBuildDir(t *testing.T) {
|
||||||
|
|||||||
@@ -1,136 +0,0 @@
|
|||||||
package deploy_test
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"database/sql"
|
|
||||||
"log/slog"
|
|
||||||
"os"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
|
||||||
"sneak.berlin/go/upaas/internal/docker"
|
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
|
||||||
"sneak.berlin/go/upaas/internal/service/deploy"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestBuildContainerOptionsUsesImageID(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
db := database.NewTestDatabase(t)
|
|
||||||
|
|
||||||
app := models.NewApp(db)
|
|
||||||
app.Name = "myapp"
|
|
||||||
|
|
||||||
err := app.Save(context.Background())
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to save app: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
|
||||||
svc := deploy.NewTestService(log)
|
|
||||||
|
|
||||||
const expectedImageID = docker.ImageID("sha256:abc123def456")
|
|
||||||
|
|
||||||
opts, err := svc.BuildContainerOptionsExported(context.Background(), app, expectedImageID)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("buildContainerOptions returned error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.Image != expectedImageID.String() {
|
|
||||||
t.Errorf("expected Image=%q, got %q", expectedImageID, opts.Image)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.Name != "upaas-myapp" {
|
|
||||||
t.Errorf("expected Name=%q, got %q", "upaas-myapp", opts.Name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestBuildContainerOptionsNoResourceLimits(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
db := database.NewTestDatabase(t)
|
|
||||||
|
|
||||||
app := models.NewApp(db)
|
|
||||||
app.Name = "nolimits"
|
|
||||||
|
|
||||||
err := app.Save(context.Background())
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to save app: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
|
||||||
svc := deploy.NewTestService(log)
|
|
||||||
|
|
||||||
opts, err := svc.BuildContainerOptionsExported(
|
|
||||||
context.Background(), app, docker.ImageID("test:latest"),
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("buildContainerOptions returned error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.CPULimit != 0 {
|
|
||||||
t.Errorf("expected CPULimit=0, got %v", opts.CPULimit)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.MemoryLimit != 0 {
|
|
||||||
t.Errorf("expected MemoryLimit=0, got %v", opts.MemoryLimit)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestBuildContainerOptionsCPULimit(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
db := database.NewTestDatabase(t)
|
|
||||||
|
|
||||||
app := models.NewApp(db)
|
|
||||||
app.Name = "cpulimit"
|
|
||||||
app.CPULimit = sql.NullFloat64{Float64: 0.5, Valid: true}
|
|
||||||
|
|
||||||
err := app.Save(context.Background())
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to save app: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
|
||||||
svc := deploy.NewTestService(log)
|
|
||||||
|
|
||||||
opts, err := svc.BuildContainerOptionsExported(
|
|
||||||
context.Background(), app, docker.ImageID("test:latest"),
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("buildContainerOptions returned error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.CPULimit != 0.5 {
|
|
||||||
t.Errorf("expected CPULimit=0.5, got %v", opts.CPULimit)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestBuildContainerOptionsMemoryLimit(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
db := database.NewTestDatabase(t)
|
|
||||||
|
|
||||||
app := models.NewApp(db)
|
|
||||||
app.Name = "memlimit"
|
|
||||||
app.MemoryLimit = sql.NullInt64{Int64: 536870912, Valid: true} // 512m
|
|
||||||
|
|
||||||
err := app.Save(context.Background())
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to save app: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
|
||||||
svc := deploy.NewTestService(log)
|
|
||||||
|
|
||||||
opts, err := svc.BuildContainerOptionsExported(
|
|
||||||
context.Background(), app, docker.ImageID("test:latest"),
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("buildContainerOptions returned error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.MemoryLimit != 536870912 {
|
|
||||||
t.Errorf("expected MemoryLimit=536870912, got %v", opts.MemoryLimit)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -8,9 +8,8 @@ import (
|
|||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/docker"
|
"git.eeqj.de/sneak/upaas/internal/docker"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// NewTestService creates a Service with minimal dependencies for testing.
|
// NewTestService creates a Service with minimal dependencies for testing.
|
||||||
@@ -81,12 +80,3 @@ func (svc *Service) CleanupCancelledDeploy(
|
|||||||
func (svc *Service) GetBuildDirExported(appName string) string {
|
func (svc *Service) GetBuildDirExported(appName string) string {
|
||||||
return svc.GetBuildDir(appName)
|
return svc.GetBuildDir(appName)
|
||||||
}
|
}
|
||||||
|
|
||||||
// BuildContainerOptionsExported exposes buildContainerOptions for testing.
|
|
||||||
func (svc *Service) BuildContainerOptionsExported(
|
|
||||||
ctx context.Context,
|
|
||||||
app *models.App,
|
|
||||||
imageID docker.ImageID,
|
|
||||||
) (docker.CreateContainerOptions, error) {
|
|
||||||
return svc.buildContainerOptions(ctx, app, imageID)
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -15,8 +15,8 @@ import (
|
|||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
)
|
)
|
||||||
|
|
||||||
// HTTP client timeout.
|
// HTTP client timeout.
|
||||||
|
|||||||
@@ -1,10 +0,0 @@
|
|||||||
package webhook
|
|
||||||
|
|
||||||
// UnparsedURL is a URL stored as a plain string without parsing.
|
|
||||||
// Use this instead of string when the value is known to be a URL
|
|
||||||
// but should not be parsed into a net/url.URL (e.g. webhook URLs,
|
|
||||||
// compare URLs from external payloads).
|
|
||||||
type UnparsedURL string
|
|
||||||
|
|
||||||
// String implements the fmt.Stringer interface.
|
|
||||||
func (u UnparsedURL) String() string { return string(u) }
|
|
||||||
@@ -10,11 +10,10 @@ import (
|
|||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
||||||
"sneak.berlin/go/upaas/internal/service/deploy"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// ServiceParams contains dependencies for Service.
|
// ServiceParams contains dependencies for Service.
|
||||||
@@ -51,12 +50,12 @@ type GiteaPushPayload struct {
|
|||||||
Ref string `json:"ref"`
|
Ref string `json:"ref"`
|
||||||
Before string `json:"before"`
|
Before string `json:"before"`
|
||||||
After string `json:"after"`
|
After string `json:"after"`
|
||||||
CompareURL UnparsedURL `json:"compare_url"`
|
CompareURL string `json:"compare_url"`
|
||||||
Repository struct {
|
Repository struct {
|
||||||
FullName string `json:"full_name"`
|
FullName string `json:"full_name"`
|
||||||
CloneURL UnparsedURL `json:"clone_url"`
|
CloneURL string `json:"clone_url"`
|
||||||
SSHURL string `json:"ssh_url"`
|
SSHURL string `json:"ssh_url"`
|
||||||
HTMLURL UnparsedURL `json:"html_url"`
|
HTMLURL string `json:"html_url"`
|
||||||
} `json:"repository"`
|
} `json:"repository"`
|
||||||
Pusher struct {
|
Pusher struct {
|
||||||
Username string `json:"username"`
|
Username string `json:"username"`
|
||||||
@@ -64,7 +63,7 @@ type GiteaPushPayload struct {
|
|||||||
} `json:"pusher"`
|
} `json:"pusher"`
|
||||||
Commits []struct {
|
Commits []struct {
|
||||||
ID string `json:"id"`
|
ID string `json:"id"`
|
||||||
URL UnparsedURL `json:"url"`
|
URL string `json:"url"`
|
||||||
Message string `json:"message"`
|
Message string `json:"message"`
|
||||||
Author struct {
|
Author struct {
|
||||||
Name string `json:"name"`
|
Name string `json:"name"`
|
||||||
@@ -105,7 +104,7 @@ func (svc *Service) HandleWebhook(
|
|||||||
event.EventType = eventType
|
event.EventType = eventType
|
||||||
event.Branch = branch
|
event.Branch = branch
|
||||||
event.CommitSHA = sql.NullString{String: commitSHA, Valid: commitSHA != ""}
|
event.CommitSHA = sql.NullString{String: commitSHA, Valid: commitSHA != ""}
|
||||||
event.CommitURL = sql.NullString{String: commitURL.String(), Valid: commitURL != ""}
|
event.CommitURL = sql.NullString{String: commitURL, Valid: commitURL != ""}
|
||||||
event.Payload = sql.NullString{String: string(payload), Valid: true}
|
event.Payload = sql.NullString{String: string(payload), Valid: true}
|
||||||
event.Matched = matched
|
event.Matched = matched
|
||||||
event.Processed = false
|
event.Processed = false
|
||||||
@@ -169,7 +168,7 @@ func extractBranch(ref string) string {
|
|||||||
|
|
||||||
// extractCommitURL extracts the commit URL from the webhook payload.
|
// extractCommitURL extracts the commit URL from the webhook payload.
|
||||||
// Prefers the URL from the head commit, falls back to constructing from repo URL.
|
// Prefers the URL from the head commit, falls back to constructing from repo URL.
|
||||||
func extractCommitURL(payload GiteaPushPayload) UnparsedURL {
|
func extractCommitURL(payload GiteaPushPayload) string {
|
||||||
// Try to find the URL from the head commit (matching After SHA)
|
// Try to find the URL from the head commit (matching After SHA)
|
||||||
for _, commit := range payload.Commits {
|
for _, commit := range payload.Commits {
|
||||||
if commit.ID == payload.After && commit.URL != "" {
|
if commit.ID == payload.After && commit.URL != "" {
|
||||||
@@ -179,7 +178,7 @@ func extractCommitURL(payload GiteaPushPayload) UnparsedURL {
|
|||||||
|
|
||||||
// Fall back to constructing URL from repo HTML URL
|
// Fall back to constructing URL from repo HTML URL
|
||||||
if payload.Repository.HTMLURL != "" && payload.After != "" {
|
if payload.Repository.HTMLURL != "" && payload.After != "" {
|
||||||
return UnparsedURL(payload.Repository.HTMLURL.String() + "/commit/" + payload.After)
|
return payload.Repository.HTMLURL + "/commit/" + payload.After
|
||||||
}
|
}
|
||||||
|
|
||||||
return ""
|
return ""
|
||||||
|
|||||||
@@ -12,15 +12,15 @@ import (
|
|||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
|
|
||||||
"sneak.berlin/go/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"sneak.berlin/go/upaas/internal/database"
|
"git.eeqj.de/sneak/upaas/internal/database"
|
||||||
"sneak.berlin/go/upaas/internal/docker"
|
"git.eeqj.de/sneak/upaas/internal/docker"
|
||||||
"sneak.berlin/go/upaas/internal/globals"
|
"git.eeqj.de/sneak/upaas/internal/globals"
|
||||||
"sneak.berlin/go/upaas/internal/logger"
|
"git.eeqj.de/sneak/upaas/internal/logger"
|
||||||
"sneak.berlin/go/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
"sneak.berlin/go/upaas/internal/service/deploy"
|
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
||||||
"sneak.berlin/go/upaas/internal/service/notify"
|
"git.eeqj.de/sneak/upaas/internal/service/notify"
|
||||||
"sneak.berlin/go/upaas/internal/service/webhook"
|
"git.eeqj.de/sneak/upaas/internal/service/webhook"
|
||||||
)
|
)
|
||||||
|
|
||||||
type testDeps struct {
|
type testDeps struct {
|
||||||
|
|||||||
@@ -4,9 +4,9 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/ssh"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
"sneak.berlin/go/upaas/internal/ssh"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestGenerateKeyPair(t *testing.T) {
|
func TestGenerateKeyPair(t *testing.T) {
|
||||||
|
|||||||
3047
static/js/alpine.min.js
vendored
3047
static/js/alpine.min.js
vendored
File diff suppressed because one or more lines are too long
@@ -1,317 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - App Detail Page Component
|
|
||||||
*
|
|
||||||
* Handles the single-app view: status polling, container logs,
|
|
||||||
* build logs, and recent deployments list.
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
// ============================================
|
|
||||||
// Environment Variable Editor Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("envVarEditor", (appId) => ({
|
|
||||||
vars: [],
|
|
||||||
editIdx: -1,
|
|
||||||
editKey: "",
|
|
||||||
editVal: "",
|
|
||||||
appId: appId,
|
|
||||||
|
|
||||||
init() {
|
|
||||||
this.vars = Array.from(this.$el.querySelectorAll(".env-init")).map(
|
|
||||||
(span) => ({
|
|
||||||
key: span.dataset.key,
|
|
||||||
value: span.dataset.value,
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
},
|
|
||||||
|
|
||||||
startEdit(i) {
|
|
||||||
this.editIdx = i;
|
|
||||||
this.editKey = this.vars[i].key;
|
|
||||||
this.editVal = this.vars[i].value;
|
|
||||||
},
|
|
||||||
|
|
||||||
saveEdit(i) {
|
|
||||||
this.vars[i] = { key: this.editKey, value: this.editVal };
|
|
||||||
this.editIdx = -1;
|
|
||||||
this.submitAll();
|
|
||||||
},
|
|
||||||
|
|
||||||
removeVar(i) {
|
|
||||||
if (!window.confirm("Delete this environment variable?")) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
this.vars.splice(i, 1);
|
|
||||||
this.submitAll();
|
|
||||||
},
|
|
||||||
|
|
||||||
addVar(keyEl, valEl) {
|
|
||||||
const k = keyEl.value.trim();
|
|
||||||
const v = valEl.value.trim();
|
|
||||||
|
|
||||||
if (!k) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
this.vars.push({ key: k, value: v });
|
|
||||||
this.submitAll();
|
|
||||||
},
|
|
||||||
|
|
||||||
submitAll() {
|
|
||||||
const csrfInput = this.$el.querySelector(
|
|
||||||
'input[name="gorilla.csrf.Token"]',
|
|
||||||
);
|
|
||||||
const csrfToken = csrfInput ? csrfInput.value : "";
|
|
||||||
|
|
||||||
fetch("/apps/" + this.appId + "/env", {
|
|
||||||
method: "POST",
|
|
||||||
headers: {
|
|
||||||
"Content-Type": "application/json",
|
|
||||||
"X-CSRF-Token": csrfToken,
|
|
||||||
},
|
|
||||||
body: JSON.stringify(
|
|
||||||
this.vars.map((e) => ({ key: e.key, value: e.value })),
|
|
||||||
),
|
|
||||||
})
|
|
||||||
.then((res) => {
|
|
||||||
if (res.ok) {
|
|
||||||
window.location.reload();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
res.json()
|
|
||||||
.then((data) => {
|
|
||||||
window.alert(
|
|
||||||
data.error ||
|
|
||||||
"Failed to save environment variables.",
|
|
||||||
);
|
|
||||||
})
|
|
||||||
.catch(() => {
|
|
||||||
window.alert(
|
|
||||||
"Failed to save environment variables.",
|
|
||||||
);
|
|
||||||
});
|
|
||||||
})
|
|
||||||
.catch(() => {
|
|
||||||
window.alert(
|
|
||||||
"Network error: could not save environment variables.",
|
|
||||||
);
|
|
||||||
});
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// App Detail Page Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("appDetail", (config) => ({
|
|
||||||
appId: config.appId,
|
|
||||||
currentDeploymentId: config.initialDeploymentId,
|
|
||||||
appStatus: config.initialStatus || "unknown",
|
|
||||||
containerLogs: "Loading container logs...",
|
|
||||||
containerStatus: "unknown",
|
|
||||||
buildLogs: config.initialDeploymentId
|
|
||||||
? "Loading build logs..."
|
|
||||||
: "No deployments yet",
|
|
||||||
buildStatus: config.initialBuildStatus || "unknown",
|
|
||||||
showBuildLogs: !!config.initialDeploymentId,
|
|
||||||
deploying: false,
|
|
||||||
deployments: [],
|
|
||||||
// Track whether user wants auto-scroll (per log pane)
|
|
||||||
_containerAutoScroll: true,
|
|
||||||
_buildAutoScroll: true,
|
|
||||||
_pollTimer: null,
|
|
||||||
|
|
||||||
init() {
|
|
||||||
this.deploying = Alpine.store("utils").isDeploying(this.appStatus);
|
|
||||||
this.fetchAll();
|
|
||||||
this._schedulePoll();
|
|
||||||
|
|
||||||
// Set up scroll listeners after DOM is ready
|
|
||||||
this.$nextTick(() => {
|
|
||||||
this._initScrollTracking(
|
|
||||||
this.$refs.containerLogsWrapper,
|
|
||||||
"_containerAutoScroll",
|
|
||||||
);
|
|
||||||
this._initScrollTracking(
|
|
||||||
this.$refs.buildLogsWrapper,
|
|
||||||
"_buildAutoScroll",
|
|
||||||
);
|
|
||||||
});
|
|
||||||
},
|
|
||||||
|
|
||||||
_schedulePoll() {
|
|
||||||
if (this._pollTimer) clearTimeout(this._pollTimer);
|
|
||||||
const interval = Alpine.store("utils").isDeploying(this.appStatus)
|
|
||||||
? 1000
|
|
||||||
: 10000;
|
|
||||||
this._pollTimer = setTimeout(() => {
|
|
||||||
this.fetchAll();
|
|
||||||
this._schedulePoll();
|
|
||||||
}, interval);
|
|
||||||
},
|
|
||||||
|
|
||||||
_initScrollTracking(el, flag) {
|
|
||||||
if (!el) return;
|
|
||||||
el.addEventListener(
|
|
||||||
"scroll",
|
|
||||||
() => {
|
|
||||||
this[flag] = Alpine.store("utils").isScrolledToBottom(el);
|
|
||||||
},
|
|
||||||
{ passive: true },
|
|
||||||
);
|
|
||||||
},
|
|
||||||
|
|
||||||
fetchAll() {
|
|
||||||
this.fetchAppStatus();
|
|
||||||
// Only fetch logs when the respective pane is visible
|
|
||||||
if (
|
|
||||||
this.$refs.containerLogsWrapper &&
|
|
||||||
this._isElementVisible(this.$refs.containerLogsWrapper)
|
|
||||||
) {
|
|
||||||
this.fetchContainerLogs();
|
|
||||||
}
|
|
||||||
if (
|
|
||||||
this.showBuildLogs &&
|
|
||||||
this.$refs.buildLogsWrapper &&
|
|
||||||
this._isElementVisible(this.$refs.buildLogsWrapper)
|
|
||||||
) {
|
|
||||||
this.fetchBuildLogs();
|
|
||||||
}
|
|
||||||
this.fetchRecentDeployments();
|
|
||||||
},
|
|
||||||
|
|
||||||
_isElementVisible(el) {
|
|
||||||
if (!el) return false;
|
|
||||||
// Check if element is in viewport (roughly)
|
|
||||||
const rect = el.getBoundingClientRect();
|
|
||||||
return rect.bottom > 0 && rect.top < window.innerHeight;
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchAppStatus() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(`/apps/${this.appId}/status`);
|
|
||||||
const data = await res.json();
|
|
||||||
const wasDeploying = this.deploying;
|
|
||||||
this.appStatus = data.status;
|
|
||||||
this.deploying = Alpine.store("utils").isDeploying(data.status);
|
|
||||||
|
|
||||||
// Re-schedule polling when deployment state changes
|
|
||||||
if (this.deploying !== wasDeploying) {
|
|
||||||
this._schedulePoll();
|
|
||||||
}
|
|
||||||
|
|
||||||
if (
|
|
||||||
data.latestDeploymentID &&
|
|
||||||
data.latestDeploymentID !== this.currentDeploymentId
|
|
||||||
) {
|
|
||||||
this.currentDeploymentId = data.latestDeploymentID;
|
|
||||||
this.showBuildLogs = true;
|
|
||||||
this.fetchBuildLogs();
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
console.error("Status fetch error:", err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchContainerLogs() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(`/apps/${this.appId}/container-logs`);
|
|
||||||
const data = await res.json();
|
|
||||||
const newLogs = data.logs || "No logs available";
|
|
||||||
const changed = newLogs !== this.containerLogs;
|
|
||||||
this.containerLogs = newLogs;
|
|
||||||
this.containerStatus = data.status;
|
|
||||||
if (changed && this._containerAutoScroll) {
|
|
||||||
this.$nextTick(() => {
|
|
||||||
Alpine.store("utils").scrollToBottom(
|
|
||||||
this.$refs.containerLogsWrapper,
|
|
||||||
);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
this.containerLogs = "Failed to fetch logs";
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchBuildLogs() {
|
|
||||||
if (!this.currentDeploymentId) return;
|
|
||||||
try {
|
|
||||||
const res = await fetch(
|
|
||||||
`/apps/${this.appId}/deployments/${this.currentDeploymentId}/logs`,
|
|
||||||
);
|
|
||||||
const data = await res.json();
|
|
||||||
const newLogs = data.logs || "No build logs available";
|
|
||||||
const changed = newLogs !== this.buildLogs;
|
|
||||||
this.buildLogs = newLogs;
|
|
||||||
this.buildStatus = data.status;
|
|
||||||
if (changed && this._buildAutoScroll) {
|
|
||||||
this.$nextTick(() => {
|
|
||||||
Alpine.store("utils").scrollToBottom(
|
|
||||||
this.$refs.buildLogsWrapper,
|
|
||||||
);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
this.buildLogs = "Failed to fetch logs";
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchRecentDeployments() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(
|
|
||||||
`/apps/${this.appId}/recent-deployments`,
|
|
||||||
);
|
|
||||||
const data = await res.json();
|
|
||||||
this.deployments = data.deployments || [];
|
|
||||||
} catch (err) {
|
|
||||||
console.error("Deployments fetch error:", err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
submitDeploy() {
|
|
||||||
this.deploying = true;
|
|
||||||
},
|
|
||||||
|
|
||||||
get statusBadgeClass() {
|
|
||||||
return Alpine.store("utils").statusBadgeClass(this.appStatus);
|
|
||||||
},
|
|
||||||
|
|
||||||
get statusLabel() {
|
|
||||||
return Alpine.store("utils").statusLabel(this.appStatus);
|
|
||||||
},
|
|
||||||
|
|
||||||
get containerStatusBadgeClass() {
|
|
||||||
return (
|
|
||||||
Alpine.store("utils").statusBadgeClass(this.containerStatus) +
|
|
||||||
" text-xs"
|
|
||||||
);
|
|
||||||
},
|
|
||||||
|
|
||||||
get containerStatusLabel() {
|
|
||||||
return Alpine.store("utils").statusLabel(this.containerStatus);
|
|
||||||
},
|
|
||||||
|
|
||||||
get buildStatusBadgeClass() {
|
|
||||||
return (
|
|
||||||
Alpine.store("utils").statusBadgeClass(this.buildStatus) +
|
|
||||||
" text-xs"
|
|
||||||
);
|
|
||||||
},
|
|
||||||
|
|
||||||
get buildStatusLabel() {
|
|
||||||
return Alpine.store("utils").statusLabel(this.buildStatus);
|
|
||||||
},
|
|
||||||
|
|
||||||
deploymentStatusClass(status) {
|
|
||||||
return Alpine.store("utils").statusBadgeClass(status);
|
|
||||||
},
|
|
||||||
|
|
||||||
deploymentStatusLabel(status) {
|
|
||||||
return Alpine.store("utils").statusLabel(status);
|
|
||||||
},
|
|
||||||
|
|
||||||
formatTime(isoTime) {
|
|
||||||
return Alpine.store("utils").formatRelativeTime(isoTime);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
581
static/js/app.js
Normal file
581
static/js/app.js
Normal file
@@ -0,0 +1,581 @@
|
|||||||
|
/**
|
||||||
|
* upaas - Frontend JavaScript with Alpine.js
|
||||||
|
*/
|
||||||
|
|
||||||
|
document.addEventListener("alpine:init", () => {
|
||||||
|
// ============================================
|
||||||
|
// Global Utilities Store
|
||||||
|
// ============================================
|
||||||
|
Alpine.store("utils", {
|
||||||
|
/**
|
||||||
|
* Format a date string as relative time (e.g., "5 minutes ago")
|
||||||
|
*/
|
||||||
|
formatRelativeTime(dateStr) {
|
||||||
|
if (!dateStr) return "";
|
||||||
|
const date = new Date(dateStr);
|
||||||
|
const now = new Date();
|
||||||
|
const diffMs = now - date;
|
||||||
|
const diffSec = Math.floor(diffMs / 1000);
|
||||||
|
const diffMin = Math.floor(diffSec / 60);
|
||||||
|
const diffHour = Math.floor(diffMin / 60);
|
||||||
|
const diffDay = Math.floor(diffHour / 24);
|
||||||
|
|
||||||
|
if (diffSec < 60) return "just now";
|
||||||
|
if (diffMin < 60)
|
||||||
|
return diffMin + (diffMin === 1 ? " minute ago" : " minutes ago");
|
||||||
|
if (diffHour < 24)
|
||||||
|
return diffHour + (diffHour === 1 ? " hour ago" : " hours ago");
|
||||||
|
if (diffDay < 7)
|
||||||
|
return diffDay + (diffDay === 1 ? " day ago" : " days ago");
|
||||||
|
return date.toLocaleDateString();
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the badge class for a given status
|
||||||
|
*/
|
||||||
|
statusBadgeClass(status) {
|
||||||
|
if (status === "running" || status === "success") return "badge-success";
|
||||||
|
if (status === "building" || status === "deploying")
|
||||||
|
return "badge-warning";
|
||||||
|
if (status === "failed" || status === "error") return "badge-error";
|
||||||
|
return "badge-neutral";
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Format status for display (capitalize first letter)
|
||||||
|
*/
|
||||||
|
statusLabel(status) {
|
||||||
|
if (!status) return "";
|
||||||
|
return status.charAt(0).toUpperCase() + status.slice(1);
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Check if status indicates active deployment
|
||||||
|
*/
|
||||||
|
isDeploying(status) {
|
||||||
|
return status === "building" || status === "deploying";
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Scroll an element to the bottom
|
||||||
|
*/
|
||||||
|
scrollToBottom(el) {
|
||||||
|
if (el) {
|
||||||
|
requestAnimationFrame(() => {
|
||||||
|
el.scrollTop = el.scrollHeight;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Check if a scrollable element is at (or near) the bottom.
|
||||||
|
* Tolerance of 30px accounts for rounding and partial lines.
|
||||||
|
*/
|
||||||
|
isScrolledToBottom(el, tolerance = 30) {
|
||||||
|
if (!el) return true;
|
||||||
|
return el.scrollHeight - el.scrollTop - el.clientHeight <= tolerance;
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Copy text to clipboard
|
||||||
|
*/
|
||||||
|
async copyToClipboard(text, button) {
|
||||||
|
try {
|
||||||
|
await navigator.clipboard.writeText(text);
|
||||||
|
return true;
|
||||||
|
} catch (err) {
|
||||||
|
// Fallback for older browsers
|
||||||
|
const textArea = document.createElement("textarea");
|
||||||
|
textArea.value = text;
|
||||||
|
textArea.style.position = "fixed";
|
||||||
|
textArea.style.left = "-9999px";
|
||||||
|
document.body.appendChild(textArea);
|
||||||
|
textArea.select();
|
||||||
|
try {
|
||||||
|
document.execCommand("copy");
|
||||||
|
document.body.removeChild(textArea);
|
||||||
|
return true;
|
||||||
|
} catch (e) {
|
||||||
|
document.body.removeChild(textArea);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Copy Button Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("copyButton", (targetId) => ({
|
||||||
|
copied: false,
|
||||||
|
async copy() {
|
||||||
|
const target = document.getElementById(targetId);
|
||||||
|
if (!target) return;
|
||||||
|
const text = target.textContent || target.value;
|
||||||
|
const success = await Alpine.store("utils").copyToClipboard(text);
|
||||||
|
if (success) {
|
||||||
|
this.copied = true;
|
||||||
|
setTimeout(() => {
|
||||||
|
this.copied = false;
|
||||||
|
}, 2000);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Confirm Action Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("confirmAction", (message) => ({
|
||||||
|
confirm(event) {
|
||||||
|
if (!window.confirm(message)) {
|
||||||
|
event.preventDefault();
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Auto-dismiss Alert Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("autoDismiss", (delay = 5000) => ({
|
||||||
|
show: true,
|
||||||
|
init() {
|
||||||
|
setTimeout(() => {
|
||||||
|
this.dismiss();
|
||||||
|
}, delay);
|
||||||
|
},
|
||||||
|
dismiss() {
|
||||||
|
this.show = false;
|
||||||
|
setTimeout(() => {
|
||||||
|
this.$el.remove();
|
||||||
|
}, 300);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Relative Time Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("relativeTime", (isoTime) => ({
|
||||||
|
display: "",
|
||||||
|
init() {
|
||||||
|
this.update();
|
||||||
|
// Update every minute
|
||||||
|
setInterval(() => this.update(), 60000);
|
||||||
|
},
|
||||||
|
update() {
|
||||||
|
this.display = Alpine.store("utils").formatRelativeTime(isoTime);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// App Detail Page Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("appDetail", (config) => ({
|
||||||
|
appId: config.appId,
|
||||||
|
currentDeploymentId: config.initialDeploymentId,
|
||||||
|
appStatus: config.initialStatus || "unknown",
|
||||||
|
containerLogs: "Loading container logs...",
|
||||||
|
containerStatus: "unknown",
|
||||||
|
buildLogs: config.initialDeploymentId
|
||||||
|
? "Loading build logs..."
|
||||||
|
: "No deployments yet",
|
||||||
|
buildStatus: config.initialBuildStatus || "unknown",
|
||||||
|
showBuildLogs: !!config.initialDeploymentId,
|
||||||
|
deploying: false,
|
||||||
|
deployments: [],
|
||||||
|
// Track whether user wants auto-scroll (per log pane)
|
||||||
|
_containerAutoScroll: true,
|
||||||
|
_buildAutoScroll: true,
|
||||||
|
_pollTimer: null,
|
||||||
|
|
||||||
|
init() {
|
||||||
|
this.deploying = Alpine.store("utils").isDeploying(this.appStatus);
|
||||||
|
this.fetchAll();
|
||||||
|
this._schedulePoll();
|
||||||
|
|
||||||
|
// Set up scroll listeners after DOM is ready
|
||||||
|
this.$nextTick(() => {
|
||||||
|
this._initScrollTracking(this.$refs.containerLogsWrapper, '_containerAutoScroll');
|
||||||
|
this._initScrollTracking(this.$refs.buildLogsWrapper, '_buildAutoScroll');
|
||||||
|
});
|
||||||
|
},
|
||||||
|
|
||||||
|
_schedulePoll() {
|
||||||
|
if (this._pollTimer) clearTimeout(this._pollTimer);
|
||||||
|
const interval = Alpine.store("utils").isDeploying(this.appStatus) ? 1000 : 10000;
|
||||||
|
this._pollTimer = setTimeout(() => {
|
||||||
|
this.fetchAll();
|
||||||
|
this._schedulePoll();
|
||||||
|
}, interval);
|
||||||
|
},
|
||||||
|
|
||||||
|
_initScrollTracking(el, flag) {
|
||||||
|
if (!el) return;
|
||||||
|
el.addEventListener('scroll', () => {
|
||||||
|
this[flag] = Alpine.store("utils").isScrolledToBottom(el);
|
||||||
|
}, { passive: true });
|
||||||
|
},
|
||||||
|
|
||||||
|
fetchAll() {
|
||||||
|
this.fetchAppStatus();
|
||||||
|
// Only fetch logs when the respective pane is visible
|
||||||
|
if (this.$refs.containerLogsWrapper && this._isElementVisible(this.$refs.containerLogsWrapper)) {
|
||||||
|
this.fetchContainerLogs();
|
||||||
|
}
|
||||||
|
if (this.showBuildLogs && this.$refs.buildLogsWrapper && this._isElementVisible(this.$refs.buildLogsWrapper)) {
|
||||||
|
this.fetchBuildLogs();
|
||||||
|
}
|
||||||
|
this.fetchRecentDeployments();
|
||||||
|
},
|
||||||
|
|
||||||
|
_isElementVisible(el) {
|
||||||
|
if (!el) return false;
|
||||||
|
// Check if element is in viewport (roughly)
|
||||||
|
const rect = el.getBoundingClientRect();
|
||||||
|
return rect.bottom > 0 && rect.top < window.innerHeight;
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchAppStatus() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(`/apps/${this.appId}/status`);
|
||||||
|
const data = await res.json();
|
||||||
|
const wasDeploying = this.deploying;
|
||||||
|
this.appStatus = data.status;
|
||||||
|
this.deploying = Alpine.store("utils").isDeploying(data.status);
|
||||||
|
|
||||||
|
// Re-schedule polling when deployment state changes
|
||||||
|
if (this.deploying !== wasDeploying) {
|
||||||
|
this._schedulePoll();
|
||||||
|
}
|
||||||
|
|
||||||
|
if (
|
||||||
|
data.latestDeploymentID &&
|
||||||
|
data.latestDeploymentID !== this.currentDeploymentId
|
||||||
|
) {
|
||||||
|
this.currentDeploymentId = data.latestDeploymentID;
|
||||||
|
this.showBuildLogs = true;
|
||||||
|
this.fetchBuildLogs();
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
console.error("Status fetch error:", err);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchContainerLogs() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(`/apps/${this.appId}/container-logs`);
|
||||||
|
const data = await res.json();
|
||||||
|
const newLogs = data.logs || "No logs available";
|
||||||
|
const changed = newLogs !== this.containerLogs;
|
||||||
|
this.containerLogs = newLogs;
|
||||||
|
this.containerStatus = data.status;
|
||||||
|
if (changed && this._containerAutoScroll) {
|
||||||
|
this.$nextTick(() => {
|
||||||
|
Alpine.store("utils").scrollToBottom(this.$refs.containerLogsWrapper);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
this.containerLogs = "Failed to fetch logs";
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchBuildLogs() {
|
||||||
|
if (!this.currentDeploymentId) return;
|
||||||
|
try {
|
||||||
|
const res = await fetch(
|
||||||
|
`/apps/${this.appId}/deployments/${this.currentDeploymentId}/logs`,
|
||||||
|
);
|
||||||
|
const data = await res.json();
|
||||||
|
const newLogs = data.logs || "No build logs available";
|
||||||
|
const changed = newLogs !== this.buildLogs;
|
||||||
|
this.buildLogs = newLogs;
|
||||||
|
this.buildStatus = data.status;
|
||||||
|
if (changed && this._buildAutoScroll) {
|
||||||
|
this.$nextTick(() => {
|
||||||
|
Alpine.store("utils").scrollToBottom(this.$refs.buildLogsWrapper);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
this.buildLogs = "Failed to fetch logs";
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchRecentDeployments() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(`/apps/${this.appId}/recent-deployments`);
|
||||||
|
const data = await res.json();
|
||||||
|
this.deployments = data.deployments || [];
|
||||||
|
} catch (err) {
|
||||||
|
console.error("Deployments fetch error:", err);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
submitDeploy() {
|
||||||
|
this.deploying = true;
|
||||||
|
},
|
||||||
|
|
||||||
|
get statusBadgeClass() {
|
||||||
|
return Alpine.store("utils").statusBadgeClass(this.appStatus);
|
||||||
|
},
|
||||||
|
|
||||||
|
get statusLabel() {
|
||||||
|
return Alpine.store("utils").statusLabel(this.appStatus);
|
||||||
|
},
|
||||||
|
|
||||||
|
get containerStatusBadgeClass() {
|
||||||
|
return (
|
||||||
|
Alpine.store("utils").statusBadgeClass(this.containerStatus) +
|
||||||
|
" text-xs"
|
||||||
|
);
|
||||||
|
},
|
||||||
|
|
||||||
|
get containerStatusLabel() {
|
||||||
|
return Alpine.store("utils").statusLabel(this.containerStatus);
|
||||||
|
},
|
||||||
|
|
||||||
|
get buildStatusBadgeClass() {
|
||||||
|
return (
|
||||||
|
Alpine.store("utils").statusBadgeClass(this.buildStatus) + " text-xs"
|
||||||
|
);
|
||||||
|
},
|
||||||
|
|
||||||
|
get buildStatusLabel() {
|
||||||
|
return Alpine.store("utils").statusLabel(this.buildStatus);
|
||||||
|
},
|
||||||
|
|
||||||
|
deploymentStatusClass(status) {
|
||||||
|
return Alpine.store("utils").statusBadgeClass(status);
|
||||||
|
},
|
||||||
|
|
||||||
|
deploymentStatusLabel(status) {
|
||||||
|
return Alpine.store("utils").statusLabel(status);
|
||||||
|
},
|
||||||
|
|
||||||
|
formatTime(isoTime) {
|
||||||
|
return Alpine.store("utils").formatRelativeTime(isoTime);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Deployment Card Component (for individual deployment cards)
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("deploymentCard", (config) => ({
|
||||||
|
appId: config.appId,
|
||||||
|
deploymentId: config.deploymentId,
|
||||||
|
logs: "",
|
||||||
|
status: config.status || "",
|
||||||
|
pollInterval: null,
|
||||||
|
_autoScroll: true,
|
||||||
|
|
||||||
|
init() {
|
||||||
|
// Read initial logs from script tag (avoids escaping issues)
|
||||||
|
const initialLogsEl = this.$el.querySelector(".initial-logs");
|
||||||
|
this.logs = initialLogsEl?.dataset.logs || "Loading...";
|
||||||
|
|
||||||
|
// Set up scroll tracking
|
||||||
|
this.$nextTick(() => {
|
||||||
|
const wrapper = this.$refs.logsWrapper;
|
||||||
|
if (wrapper) {
|
||||||
|
wrapper.addEventListener('scroll', () => {
|
||||||
|
this._autoScroll = Alpine.store("utils").isScrolledToBottom(wrapper);
|
||||||
|
}, { passive: true });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// Only poll if deployment is in progress
|
||||||
|
if (Alpine.store("utils").isDeploying(this.status)) {
|
||||||
|
this.fetchLogs();
|
||||||
|
this.pollInterval = setInterval(() => this.fetchLogs(), 1000);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
destroy() {
|
||||||
|
if (this.pollInterval) {
|
||||||
|
clearInterval(this.pollInterval);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchLogs() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(
|
||||||
|
`/apps/${this.appId}/deployments/${this.deploymentId}/logs`,
|
||||||
|
);
|
||||||
|
const data = await res.json();
|
||||||
|
const newLogs = data.logs || "No logs available";
|
||||||
|
const logsChanged = newLogs !== this.logs;
|
||||||
|
this.logs = newLogs;
|
||||||
|
this.status = data.status;
|
||||||
|
|
||||||
|
// Scroll to bottom only when content changes AND user hasn't scrolled up
|
||||||
|
if (logsChanged && this._autoScroll) {
|
||||||
|
this.$nextTick(() => {
|
||||||
|
Alpine.store("utils").scrollToBottom(this.$refs.logsWrapper);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// Stop polling if deployment is done
|
||||||
|
if (!Alpine.store("utils").isDeploying(data.status)) {
|
||||||
|
if (this.pollInterval) {
|
||||||
|
clearInterval(this.pollInterval);
|
||||||
|
this.pollInterval = null;
|
||||||
|
}
|
||||||
|
// Reload page to show final state with duration etc
|
||||||
|
window.location.reload();
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
console.error("Logs fetch error:", err);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
get statusBadgeClass() {
|
||||||
|
return Alpine.store("utils").statusBadgeClass(this.status);
|
||||||
|
},
|
||||||
|
|
||||||
|
get statusLabel() {
|
||||||
|
return Alpine.store("utils").statusLabel(this.status);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Deployments History Page Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("deploymentsPage", (config) => ({
|
||||||
|
appId: config.appId,
|
||||||
|
currentDeploymentId: null,
|
||||||
|
isDeploying: false,
|
||||||
|
|
||||||
|
init() {
|
||||||
|
// Check for in-progress deployments on page load
|
||||||
|
const inProgressCard = document.querySelector(
|
||||||
|
'[data-status="building"], [data-status="deploying"]',
|
||||||
|
);
|
||||||
|
if (inProgressCard) {
|
||||||
|
this.currentDeploymentId = parseInt(
|
||||||
|
inProgressCard.getAttribute("data-deployment-id"),
|
||||||
|
10,
|
||||||
|
);
|
||||||
|
this.isDeploying = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
this.fetchAppStatus();
|
||||||
|
this._scheduleStatusPoll();
|
||||||
|
},
|
||||||
|
|
||||||
|
_statusPollTimer: null,
|
||||||
|
|
||||||
|
_scheduleStatusPoll() {
|
||||||
|
if (this._statusPollTimer) clearTimeout(this._statusPollTimer);
|
||||||
|
const interval = this.isDeploying ? 1000 : 10000;
|
||||||
|
this._statusPollTimer = setTimeout(() => {
|
||||||
|
this.fetchAppStatus();
|
||||||
|
this._scheduleStatusPoll();
|
||||||
|
}, interval);
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchAppStatus() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(`/apps/${this.appId}/status`);
|
||||||
|
const data = await res.json();
|
||||||
|
// Use deployment status, not app status - it's more reliable during transitions
|
||||||
|
const deploying = Alpine.store("utils").isDeploying(
|
||||||
|
data.latestDeploymentStatus,
|
||||||
|
);
|
||||||
|
|
||||||
|
// Detect new deployment
|
||||||
|
if (
|
||||||
|
data.latestDeploymentID &&
|
||||||
|
data.latestDeploymentID !== this.currentDeploymentId
|
||||||
|
) {
|
||||||
|
// Check if we have a card for this deployment
|
||||||
|
const hasCard = document.querySelector(
|
||||||
|
`[data-deployment-id="${data.latestDeploymentID}"]`,
|
||||||
|
);
|
||||||
|
|
||||||
|
if (deploying && !hasCard) {
|
||||||
|
// New deployment started but no card exists - reload to show it
|
||||||
|
window.location.reload();
|
||||||
|
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
this.currentDeploymentId = data.latestDeploymentID;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update deploying state based on latest deployment status
|
||||||
|
if (deploying && !this.isDeploying) {
|
||||||
|
this.isDeploying = true;
|
||||||
|
this._scheduleStatusPoll(); // Switch to fast polling
|
||||||
|
} else if (!deploying && this.isDeploying) {
|
||||||
|
// Deployment finished - reload to show final state
|
||||||
|
this.isDeploying = false;
|
||||||
|
window.location.reload();
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
console.error("Status fetch error:", err);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
submitDeploy() {
|
||||||
|
this.isDeploying = true;
|
||||||
|
},
|
||||||
|
|
||||||
|
formatTime(isoTime) {
|
||||||
|
return Alpine.store("utils").formatRelativeTime(isoTime);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Dashboard Page - Relative Time Updates
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("dashboard", () => ({
|
||||||
|
init() {
|
||||||
|
// Update relative times every minute
|
||||||
|
setInterval(() => {
|
||||||
|
this.$el.querySelectorAll("[data-time]").forEach((el) => {
|
||||||
|
const time = el.getAttribute("data-time");
|
||||||
|
if (time) {
|
||||||
|
el.textContent = Alpine.store("utils").formatRelativeTime(time);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}, 60000);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
});
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Legacy support - expose utilities globally
|
||||||
|
// ============================================
|
||||||
|
window.upaas = {
|
||||||
|
// These are kept for backwards compatibility but templates should use Alpine.js
|
||||||
|
formatRelativeTime(dateStr) {
|
||||||
|
if (!dateStr) return "";
|
||||||
|
const date = new Date(dateStr);
|
||||||
|
const now = new Date();
|
||||||
|
const diffMs = now - date;
|
||||||
|
const diffSec = Math.floor(diffMs / 1000);
|
||||||
|
const diffMin = Math.floor(diffSec / 60);
|
||||||
|
const diffHour = Math.floor(diffMin / 60);
|
||||||
|
const diffDay = Math.floor(diffHour / 24);
|
||||||
|
|
||||||
|
if (diffSec < 60) return "just now";
|
||||||
|
if (diffMin < 60)
|
||||||
|
return diffMin + (diffMin === 1 ? " minute ago" : " minutes ago");
|
||||||
|
if (diffHour < 24)
|
||||||
|
return diffHour + (diffHour === 1 ? " hour ago" : " hours ago");
|
||||||
|
if (diffDay < 7)
|
||||||
|
return diffDay + (diffDay === 1 ? " day ago" : " days ago");
|
||||||
|
return date.toLocaleDateString();
|
||||||
|
},
|
||||||
|
// Placeholder functions - templates should migrate to Alpine.js
|
||||||
|
initAppDetailPage() {},
|
||||||
|
initDeploymentsPage() {},
|
||||||
|
};
|
||||||
|
|
||||||
|
// Update relative times on page load for non-Alpine elements
|
||||||
|
document.addEventListener("DOMContentLoaded", () => {
|
||||||
|
document.querySelectorAll(".relative-time[data-time]").forEach((el) => {
|
||||||
|
const time = el.getAttribute("data-time");
|
||||||
|
if (time) {
|
||||||
|
el.textContent = window.upaas.formatRelativeTime(time);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
@@ -1,71 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - Reusable Alpine.js Components
|
|
||||||
*
|
|
||||||
* Small, self-contained components: copy button, confirm dialog,
|
|
||||||
* auto-dismiss alerts, and relative time display.
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
// ============================================
|
|
||||||
// Copy Button Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("copyButton", (targetId) => ({
|
|
||||||
copied: false,
|
|
||||||
async copy() {
|
|
||||||
const target = document.getElementById(targetId);
|
|
||||||
if (!target) return;
|
|
||||||
const text = target.textContent || target.value;
|
|
||||||
const success = await Alpine.store("utils").copyToClipboard(text);
|
|
||||||
if (success) {
|
|
||||||
this.copied = true;
|
|
||||||
setTimeout(() => {
|
|
||||||
this.copied = false;
|
|
||||||
}, 2000);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Confirm Action Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("confirmAction", (message) => ({
|
|
||||||
confirm(event) {
|
|
||||||
if (!window.confirm(message)) {
|
|
||||||
event.preventDefault();
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Auto-dismiss Alert Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("autoDismiss", (delay = 5000) => ({
|
|
||||||
show: true,
|
|
||||||
init() {
|
|
||||||
setTimeout(() => {
|
|
||||||
this.dismiss();
|
|
||||||
}, delay);
|
|
||||||
},
|
|
||||||
dismiss() {
|
|
||||||
this.show = false;
|
|
||||||
setTimeout(() => {
|
|
||||||
this.$el.remove();
|
|
||||||
}, 300);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Relative Time Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("relativeTime", (isoTime) => ({
|
|
||||||
display: "",
|
|
||||||
init() {
|
|
||||||
this.update();
|
|
||||||
// Update every minute
|
|
||||||
setInterval(() => this.update(), 60000);
|
|
||||||
},
|
|
||||||
update() {
|
|
||||||
this.display = Alpine.store("utils").formatRelativeTime(isoTime);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
@@ -1,22 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - Dashboard Page Component
|
|
||||||
*
|
|
||||||
* Periodically updates relative timestamps on the main dashboard.
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
Alpine.data("dashboard", () => ({
|
|
||||||
init() {
|
|
||||||
// Update relative times every minute
|
|
||||||
setInterval(() => {
|
|
||||||
this.$el.querySelectorAll("[data-time]").forEach((el) => {
|
|
||||||
const time = el.getAttribute("data-time");
|
|
||||||
if (time) {
|
|
||||||
el.textContent =
|
|
||||||
Alpine.store("utils").formatRelativeTime(time);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}, 60000);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
@@ -1,185 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - Deployment Components
|
|
||||||
*
|
|
||||||
* Deployment card (individual deployment log viewer) and
|
|
||||||
* deployments history page (list of all deployments).
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
// ============================================
|
|
||||||
// Deployment Card Component (for individual deployment cards)
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("deploymentCard", (config) => ({
|
|
||||||
appId: config.appId,
|
|
||||||
deploymentId: config.deploymentId,
|
|
||||||
logs: "",
|
|
||||||
status: config.status || "",
|
|
||||||
pollInterval: null,
|
|
||||||
_autoScroll: true,
|
|
||||||
|
|
||||||
init() {
|
|
||||||
// Read initial logs from script tag (avoids escaping issues)
|
|
||||||
const initialLogsEl = this.$el.querySelector(".initial-logs");
|
|
||||||
this.logs = initialLogsEl?.dataset.logs || "Loading...";
|
|
||||||
|
|
||||||
// Set up scroll tracking
|
|
||||||
this.$nextTick(() => {
|
|
||||||
const wrapper = this.$refs.logsWrapper;
|
|
||||||
if (wrapper) {
|
|
||||||
wrapper.addEventListener(
|
|
||||||
"scroll",
|
|
||||||
() => {
|
|
||||||
this._autoScroll =
|
|
||||||
Alpine.store("utils").isScrolledToBottom(
|
|
||||||
wrapper,
|
|
||||||
);
|
|
||||||
},
|
|
||||||
{ passive: true },
|
|
||||||
);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
// Only poll if deployment is in progress
|
|
||||||
if (Alpine.store("utils").isDeploying(this.status)) {
|
|
||||||
this.fetchLogs();
|
|
||||||
this.pollInterval = setInterval(() => this.fetchLogs(), 1000);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
destroy() {
|
|
||||||
if (this.pollInterval) {
|
|
||||||
clearInterval(this.pollInterval);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchLogs() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(
|
|
||||||
`/apps/${this.appId}/deployments/${this.deploymentId}/logs`,
|
|
||||||
);
|
|
||||||
const data = await res.json();
|
|
||||||
const newLogs = data.logs || "No logs available";
|
|
||||||
const logsChanged = newLogs !== this.logs;
|
|
||||||
this.logs = newLogs;
|
|
||||||
this.status = data.status;
|
|
||||||
|
|
||||||
// Scroll to bottom only when content changes AND user hasn't scrolled up
|
|
||||||
if (logsChanged && this._autoScroll) {
|
|
||||||
this.$nextTick(() => {
|
|
||||||
Alpine.store("utils").scrollToBottom(
|
|
||||||
this.$refs.logsWrapper,
|
|
||||||
);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
// Stop polling if deployment is done
|
|
||||||
if (!Alpine.store("utils").isDeploying(data.status)) {
|
|
||||||
if (this.pollInterval) {
|
|
||||||
clearInterval(this.pollInterval);
|
|
||||||
this.pollInterval = null;
|
|
||||||
}
|
|
||||||
// Reload page to show final state with duration etc
|
|
||||||
window.location.reload();
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
console.error("Logs fetch error:", err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
get statusBadgeClass() {
|
|
||||||
return Alpine.store("utils").statusBadgeClass(this.status);
|
|
||||||
},
|
|
||||||
|
|
||||||
get statusLabel() {
|
|
||||||
return Alpine.store("utils").statusLabel(this.status);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Deployments History Page Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("deploymentsPage", (config) => ({
|
|
||||||
appId: config.appId,
|
|
||||||
currentDeploymentId: null,
|
|
||||||
isDeploying: false,
|
|
||||||
|
|
||||||
init() {
|
|
||||||
// Check for in-progress deployments on page load
|
|
||||||
const inProgressCard = document.querySelector(
|
|
||||||
'[data-status="building"], [data-status="deploying"]',
|
|
||||||
);
|
|
||||||
if (inProgressCard) {
|
|
||||||
this.currentDeploymentId = parseInt(
|
|
||||||
inProgressCard.getAttribute("data-deployment-id"),
|
|
||||||
10,
|
|
||||||
);
|
|
||||||
this.isDeploying = true;
|
|
||||||
}
|
|
||||||
|
|
||||||
this.fetchAppStatus();
|
|
||||||
this._scheduleStatusPoll();
|
|
||||||
},
|
|
||||||
|
|
||||||
_statusPollTimer: null,
|
|
||||||
|
|
||||||
_scheduleStatusPoll() {
|
|
||||||
if (this._statusPollTimer) clearTimeout(this._statusPollTimer);
|
|
||||||
const interval = this.isDeploying ? 1000 : 10000;
|
|
||||||
this._statusPollTimer = setTimeout(() => {
|
|
||||||
this.fetchAppStatus();
|
|
||||||
this._scheduleStatusPoll();
|
|
||||||
}, interval);
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchAppStatus() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(`/apps/${this.appId}/status`);
|
|
||||||
const data = await res.json();
|
|
||||||
// Use deployment status, not app status - it's more reliable during transitions
|
|
||||||
const deploying = Alpine.store("utils").isDeploying(
|
|
||||||
data.latestDeploymentStatus,
|
|
||||||
);
|
|
||||||
|
|
||||||
// Detect new deployment
|
|
||||||
if (
|
|
||||||
data.latestDeploymentID &&
|
|
||||||
data.latestDeploymentID !== this.currentDeploymentId
|
|
||||||
) {
|
|
||||||
// Check if we have a card for this deployment
|
|
||||||
const hasCard = document.querySelector(
|
|
||||||
`[data-deployment-id="${data.latestDeploymentID}"]`,
|
|
||||||
);
|
|
||||||
|
|
||||||
if (deploying && !hasCard) {
|
|
||||||
// New deployment started but no card exists - reload to show it
|
|
||||||
window.location.reload();
|
|
||||||
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
this.currentDeploymentId = data.latestDeploymentID;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Update deploying state based on latest deployment status
|
|
||||||
if (deploying && !this.isDeploying) {
|
|
||||||
this.isDeploying = true;
|
|
||||||
this._scheduleStatusPoll(); // Switch to fast polling
|
|
||||||
} else if (!deploying && this.isDeploying) {
|
|
||||||
// Deployment finished - reload to show final state
|
|
||||||
this.isDeploying = false;
|
|
||||||
window.location.reload();
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
console.error("Status fetch error:", err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
submitDeploy() {
|
|
||||||
this.isDeploying = true;
|
|
||||||
},
|
|
||||||
|
|
||||||
formatTime(isoTime) {
|
|
||||||
return Alpine.store("utils").formatRelativeTime(isoTime);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
@@ -1,148 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - Global Utilities Store
|
|
||||||
*
|
|
||||||
* Shared formatting, status helpers, and clipboard utilities used across all pages.
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
Alpine.store("utils", {
|
|
||||||
/**
|
|
||||||
* Format a date string as relative time (e.g., "5 minutes ago")
|
|
||||||
*/
|
|
||||||
formatRelativeTime(dateStr) {
|
|
||||||
if (!dateStr) return "";
|
|
||||||
const date = new Date(dateStr);
|
|
||||||
const now = new Date();
|
|
||||||
const diffMs = now - date;
|
|
||||||
const diffSec = Math.floor(diffMs / 1000);
|
|
||||||
const diffMin = Math.floor(diffSec / 60);
|
|
||||||
const diffHour = Math.floor(diffMin / 60);
|
|
||||||
const diffDay = Math.floor(diffHour / 24);
|
|
||||||
|
|
||||||
if (diffSec < 60) return "just now";
|
|
||||||
if (diffMin < 60)
|
|
||||||
return (
|
|
||||||
diffMin + (diffMin === 1 ? " minute ago" : " minutes ago")
|
|
||||||
);
|
|
||||||
if (diffHour < 24)
|
|
||||||
return diffHour + (diffHour === 1 ? " hour ago" : " hours ago");
|
|
||||||
if (diffDay < 7)
|
|
||||||
return diffDay + (diffDay === 1 ? " day ago" : " days ago");
|
|
||||||
return date.toLocaleDateString();
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Get the badge class for a given status
|
|
||||||
*/
|
|
||||||
statusBadgeClass(status) {
|
|
||||||
if (status === "running" || status === "success")
|
|
||||||
return "badge-success";
|
|
||||||
if (status === "building" || status === "deploying")
|
|
||||||
return "badge-warning";
|
|
||||||
if (status === "failed" || status === "error") return "badge-error";
|
|
||||||
return "badge-neutral";
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Format status for display (capitalize first letter)
|
|
||||||
*/
|
|
||||||
statusLabel(status) {
|
|
||||||
if (!status) return "";
|
|
||||||
return status.charAt(0).toUpperCase() + status.slice(1);
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Check if status indicates active deployment
|
|
||||||
*/
|
|
||||||
isDeploying(status) {
|
|
||||||
return status === "building" || status === "deploying";
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Scroll an element to the bottom
|
|
||||||
*/
|
|
||||||
scrollToBottom(el) {
|
|
||||||
if (el) {
|
|
||||||
requestAnimationFrame(() => {
|
|
||||||
el.scrollTop = el.scrollHeight;
|
|
||||||
});
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Check if a scrollable element is at (or near) the bottom.
|
|
||||||
* Tolerance of 30px accounts for rounding and partial lines.
|
|
||||||
*/
|
|
||||||
isScrolledToBottom(el, tolerance = 30) {
|
|
||||||
if (!el) return true;
|
|
||||||
return (
|
|
||||||
el.scrollHeight - el.scrollTop - el.clientHeight <= tolerance
|
|
||||||
);
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Copy text to clipboard
|
|
||||||
*/
|
|
||||||
async copyToClipboard(text, button) {
|
|
||||||
try {
|
|
||||||
await navigator.clipboard.writeText(text);
|
|
||||||
return true;
|
|
||||||
} catch (err) {
|
|
||||||
// Fallback for older browsers
|
|
||||||
const textArea = document.createElement("textarea");
|
|
||||||
textArea.value = text;
|
|
||||||
textArea.style.position = "fixed";
|
|
||||||
textArea.style.left = "-9999px";
|
|
||||||
document.body.appendChild(textArea);
|
|
||||||
textArea.select();
|
|
||||||
try {
|
|
||||||
document.execCommand("copy");
|
|
||||||
document.body.removeChild(textArea);
|
|
||||||
return true;
|
|
||||||
} catch (e) {
|
|
||||||
document.body.removeChild(textArea);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Legacy support - expose utilities globally
|
|
||||||
// ============================================
|
|
||||||
window.upaas = {
|
|
||||||
// These are kept for backwards compatibility but templates should use Alpine.js
|
|
||||||
formatRelativeTime(dateStr) {
|
|
||||||
if (!dateStr) return "";
|
|
||||||
const date = new Date(dateStr);
|
|
||||||
const now = new Date();
|
|
||||||
const diffMs = now - date;
|
|
||||||
const diffSec = Math.floor(diffMs / 1000);
|
|
||||||
const diffMin = Math.floor(diffSec / 60);
|
|
||||||
const diffHour = Math.floor(diffMin / 60);
|
|
||||||
const diffDay = Math.floor(diffHour / 24);
|
|
||||||
|
|
||||||
if (diffSec < 60) return "just now";
|
|
||||||
if (diffMin < 60)
|
|
||||||
return diffMin + (diffMin === 1 ? " minute ago" : " minutes ago");
|
|
||||||
if (diffHour < 24)
|
|
||||||
return diffHour + (diffHour === 1 ? " hour ago" : " hours ago");
|
|
||||||
if (diffDay < 7)
|
|
||||||
return diffDay + (diffDay === 1 ? " day ago" : " days ago");
|
|
||||||
return date.toLocaleDateString();
|
|
||||||
},
|
|
||||||
// Placeholder functions - templates should migrate to Alpine.js
|
|
||||||
initAppDetailPage() {},
|
|
||||||
initDeploymentsPage() {},
|
|
||||||
};
|
|
||||||
|
|
||||||
// Update relative times on page load for non-Alpine elements
|
|
||||||
document.addEventListener("DOMContentLoaded", () => {
|
|
||||||
document.querySelectorAll(".relative-time[data-time]").forEach((el) => {
|
|
||||||
const time = el.getAttribute("data-time");
|
|
||||||
if (time) {
|
|
||||||
el.textContent = window.upaas.formatRelativeTime(time);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
});
|
|
||||||
@@ -77,10 +77,7 @@
|
|||||||
|
|
||||||
<!-- Webhook URL -->
|
<!-- Webhook URL -->
|
||||||
<div class="card p-6 mb-6">
|
<div class="card p-6 mb-6">
|
||||||
<div class="flex items-center justify-between mb-4">
|
<h2 class="section-title mb-4">Webhook URL</h2>
|
||||||
<h2 class="section-title">Webhook URL</h2>
|
|
||||||
<a href="/apps/{{.App.ID}}/webhooks" class="text-primary-600 hover:text-primary-800 text-sm">Event History</a>
|
|
||||||
</div>
|
|
||||||
<p class="text-sm text-gray-500 mb-3">Add this URL as a push webhook in your Gitea repository:</p>
|
<p class="text-sm text-gray-500 mb-3">Add this URL as a push webhook in your Gitea repository:</p>
|
||||||
<div class="copy-field" x-data="copyButton('webhook-url')">
|
<div class="copy-field" x-data="copyButton('webhook-url')">
|
||||||
<code id="webhook-url" class="copy-field-value text-xs">{{.WebhookURL}}</code>
|
<code id="webhook-url" class="copy-field-value text-xs">{{.WebhookURL}}</code>
|
||||||
@@ -101,10 +98,9 @@
|
|||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Environment Variables -->
|
<!-- Environment Variables -->
|
||||||
<div class="card p-6 mb-6" x-data="envVarEditor('{{.App.ID}}')">
|
<div class="card p-6 mb-6">
|
||||||
<h2 class="section-title mb-4">Environment Variables</h2>
|
<h2 class="section-title mb-4">Environment Variables</h2>
|
||||||
{{range .EnvVars}}<span class="env-init hidden" data-key="{{.Key}}" data-value="{{.Value}}"></span>{{end}}
|
{{if .EnvVars}}
|
||||||
<template x-if="vars.length > 0">
|
|
||||||
<div class="overflow-x-auto mb-4">
|
<div class="overflow-x-auto mb-4">
|
||||||
<table class="table">
|
<table class="table">
|
||||||
<thead class="table-header">
|
<thead class="table-header">
|
||||||
@@ -115,43 +111,47 @@
|
|||||||
</tr>
|
</tr>
|
||||||
</thead>
|
</thead>
|
||||||
<tbody class="table-body">
|
<tbody class="table-body">
|
||||||
<template x-for="(env, idx) in vars" :key="idx">
|
{{range .EnvVars}}
|
||||||
<tr>
|
<tr x-data="{ editing: false }">
|
||||||
<template x-if="editIdx !== idx">
|
<template x-if="!editing">
|
||||||
<td class="font-mono font-medium" x-text="env.key"></td>
|
<td class="font-mono font-medium">{{.Key}}</td>
|
||||||
</template>
|
</template>
|
||||||
<template x-if="editIdx !== idx">
|
<template x-if="!editing">
|
||||||
<td class="font-mono text-gray-500" x-text="env.value"></td>
|
<td class="font-mono text-gray-500">{{.Value}}</td>
|
||||||
</template>
|
</template>
|
||||||
<template x-if="editIdx !== idx">
|
<template x-if="!editing">
|
||||||
<td class="text-right">
|
<td class="text-right">
|
||||||
<button @click="startEdit(idx)" class="text-primary-600 hover:text-primary-800 text-sm mr-2">Edit</button>
|
<button @click="editing = true" class="text-primary-600 hover:text-primary-800 text-sm mr-2">Edit</button>
|
||||||
<button @click="removeVar(idx)" class="text-error-500 hover:text-error-700 text-sm">Delete</button>
|
<form method="POST" action="/apps/{{$.App.ID}}/env-vars/{{.ID}}/delete" class="inline" x-data="confirmAction('Delete this environment variable?')" @submit="confirm($event)">
|
||||||
|
{{ $.CSRFField }}
|
||||||
|
<button type="submit" class="text-error-500 hover:text-error-700 text-sm">Delete</button>
|
||||||
|
</form>
|
||||||
</td>
|
</td>
|
||||||
</template>
|
</template>
|
||||||
<template x-if="editIdx === idx">
|
<template x-if="editing">
|
||||||
<td colspan="3">
|
<td colspan="3">
|
||||||
<form @submit.prevent="saveEdit(idx)" class="flex gap-2 items-center">
|
<form method="POST" action="/apps/{{$.App.ID}}/env-vars/{{.ID}}/edit" class="flex gap-2 items-center">
|
||||||
<input type="text" x-model="editKey" required class="input flex-1 font-mono text-sm">
|
{{ $.CSRFField }}
|
||||||
<input type="text" x-model="editVal" required class="input flex-1 font-mono text-sm">
|
<input type="text" name="key" value="{{.Key}}" required class="input flex-1 font-mono text-sm">
|
||||||
|
<input type="text" name="value" value="{{.Value}}" required class="input flex-1 font-mono text-sm">
|
||||||
<button type="submit" class="btn-primary text-sm">Save</button>
|
<button type="submit" class="btn-primary text-sm">Save</button>
|
||||||
<button type="button" @click="editIdx = -1" class="text-gray-500 hover:text-gray-700 text-sm">Cancel</button>
|
<button type="button" @click="editing = false" class="text-gray-500 hover:text-gray-700 text-sm">Cancel</button>
|
||||||
</form>
|
</form>
|
||||||
<p class="text-xs text-amber-600 mt-1">⚠ Container restart needed after env var changes.</p>
|
<p class="text-xs text-amber-600 mt-1">⚠ Container restart needed after env var changes.</p>
|
||||||
</td>
|
</td>
|
||||||
</template>
|
</template>
|
||||||
</tr>
|
</tr>
|
||||||
</template>
|
{{end}}
|
||||||
</tbody>
|
</tbody>
|
||||||
</table>
|
</table>
|
||||||
</div>
|
</div>
|
||||||
</template>
|
{{end}}
|
||||||
<form @submit.prevent="addVar($refs.newKey, $refs.newVal)" class="flex flex-col sm:flex-row gap-2">
|
<form method="POST" action="/apps/{{.App.ID}}/env" class="flex flex-col sm:flex-row gap-2">
|
||||||
<input x-ref="newKey" type="text" placeholder="KEY" required class="input flex-1 font-mono text-sm">
|
{{ .CSRFField }}
|
||||||
<input x-ref="newVal" type="text" placeholder="value" required class="input flex-1 font-mono text-sm">
|
<input type="text" name="key" placeholder="KEY" required class="input flex-1 font-mono text-sm">
|
||||||
|
<input type="text" name="value" placeholder="value" required class="input flex-1 font-mono text-sm">
|
||||||
<button type="submit" class="btn-primary">Add</button>
|
<button type="submit" class="btn-primary">Add</button>
|
||||||
</form>
|
</form>
|
||||||
<div class="hidden">{{ .CSRFField }}</div>
|
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Labels -->
|
<!-- Labels -->
|
||||||
|
|||||||
@@ -114,38 +114,6 @@
|
|||||||
>
|
>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<hr class="border-gray-200">
|
|
||||||
|
|
||||||
<h3 class="text-lg font-medium text-gray-900">Resource Limits</h3>
|
|
||||||
|
|
||||||
<div class="grid grid-cols-2 gap-4">
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="cpu_limit" class="label">CPU Limit (cores)</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
id="cpu_limit"
|
|
||||||
name="cpu_limit"
|
|
||||||
value="{{if .App.CPULimit.Valid}}{{.App.CPULimit.Float64}}{{end}}"
|
|
||||||
class="input"
|
|
||||||
placeholder="e.g. 0.5, 1, 2"
|
|
||||||
>
|
|
||||||
<p class="text-sm text-gray-500 mt-1">Number of CPU cores (e.g. 0.5 = half a core)</p>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="memory_limit" class="label">Memory Limit</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
id="memory_limit"
|
|
||||||
name="memory_limit"
|
|
||||||
value="{{if .App.MemoryLimit.Valid}}{{formatMemoryBytes .App.MemoryLimit.Int64}}{{end}}"
|
|
||||||
class="input"
|
|
||||||
placeholder="e.g. 256m, 1g"
|
|
||||||
>
|
|
||||||
<p class="text-sm text-gray-500 mt-1">Memory with unit suffix (k, m, g) or plain bytes</p>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="flex justify-end gap-3 pt-4">
|
<div class="flex justify-end gap-3 pt-4">
|
||||||
<a href="/apps/{{.App.ID}}" class="btn-secondary">Cancel</a>
|
<a href="/apps/{{.App.ID}}" class="btn-secondary">Cancel</a>
|
||||||
<button type="submit" class="btn-primary">Save Changes</button>
|
<button type="submit" class="btn-primary">Save Changes</button>
|
||||||
|
|||||||
@@ -15,11 +15,7 @@
|
|||||||
</div>
|
</div>
|
||||||
{{template "footer" .}}
|
{{template "footer" .}}
|
||||||
<script defer src="/s/js/alpine.min.js"></script>
|
<script defer src="/s/js/alpine.min.js"></script>
|
||||||
<script src="/s/js/utils.js"></script>
|
<script src="/s/js/app.js"></script>
|
||||||
<script src="/s/js/components.js"></script>
|
|
||||||
<script src="/s/js/app-detail.js"></script>
|
|
||||||
<script src="/s/js/deployment.js"></script>
|
|
||||||
<script src="/s/js/dashboard.js"></script>
|
|
||||||
</body>
|
</body>
|
||||||
</html>
|
</html>
|
||||||
{{end}}
|
{{end}}
|
||||||
|
|||||||
@@ -69,7 +69,7 @@
|
|||||||
<a href="/apps/{{.App.ID}}" class="btn-text text-sm py-1 px-2">View</a>
|
<a href="/apps/{{.App.ID}}" class="btn-text text-sm py-1 px-2">View</a>
|
||||||
<a href="/apps/{{.App.ID}}/edit" class="btn-secondary text-sm py-1 px-2">Edit</a>
|
<a href="/apps/{{.App.ID}}/edit" class="btn-secondary text-sm py-1 px-2">Edit</a>
|
||||||
<form method="POST" action="/apps/{{.App.ID}}/deploy" class="inline">
|
<form method="POST" action="/apps/{{.App.ID}}/deploy" class="inline">
|
||||||
{{ $.CSRFField }}
|
{{ .CSRFField }}
|
||||||
<button type="submit" class="btn-success text-sm py-1 px-2">Deploy</button>
|
<button type="submit" class="btn-success text-sm py-1 px-2">Deploy</button>
|
||||||
</form>
|
</form>
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@@ -6,7 +6,6 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"html/template"
|
"html/template"
|
||||||
"io"
|
"io"
|
||||||
"strconv"
|
|
||||||
"sync"
|
"sync"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -24,34 +23,6 @@ var (
|
|||||||
templatesMutex sync.RWMutex
|
templatesMutex sync.RWMutex
|
||||||
)
|
)
|
||||||
|
|
||||||
// templateFuncs returns the custom template function map.
|
|
||||||
func templateFuncs() template.FuncMap {
|
|
||||||
return template.FuncMap{
|
|
||||||
"formatMemoryBytes": formatMemoryBytes,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Memory unit constants.
|
|
||||||
const (
|
|
||||||
memGigabyte = 1024 * 1024 * 1024
|
|
||||||
memMegabyte = 1024 * 1024
|
|
||||||
memKilobyte = 1024
|
|
||||||
)
|
|
||||||
|
|
||||||
// formatMemoryBytes formats bytes into a human-readable string with unit suffix.
|
|
||||||
func formatMemoryBytes(bytes int64) string {
|
|
||||||
switch {
|
|
||||||
case bytes >= memGigabyte && bytes%memGigabyte == 0:
|
|
||||||
return strconv.FormatInt(bytes/memGigabyte, 10) + "g"
|
|
||||||
case bytes >= memMegabyte && bytes%memMegabyte == 0:
|
|
||||||
return strconv.FormatInt(bytes/memMegabyte, 10) + "m"
|
|
||||||
case bytes >= memKilobyte && bytes%memKilobyte == 0:
|
|
||||||
return strconv.FormatInt(bytes/memKilobyte, 10) + "k"
|
|
||||||
default:
|
|
||||||
return strconv.FormatInt(bytes, 10)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// initTemplates parses base template and creates cloned templates for each page.
|
// initTemplates parses base template and creates cloned templates for each page.
|
||||||
func initTemplates() {
|
func initTemplates() {
|
||||||
templatesMutex.Lock()
|
templatesMutex.Lock()
|
||||||
@@ -61,10 +32,8 @@ func initTemplates() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Parse base template with shared components and custom functions
|
// Parse base template with shared components
|
||||||
baseTemplate = template.Must(
|
baseTemplate = template.Must(template.ParseFS(templatesRaw, "base.html"))
|
||||||
template.New("base.html").Funcs(templateFuncs()).ParseFS(templatesRaw, "base.html"),
|
|
||||||
)
|
|
||||||
|
|
||||||
// Pages that extend base
|
// Pages that extend base
|
||||||
pages := []string{
|
pages := []string{
|
||||||
@@ -75,7 +44,6 @@ func initTemplates() {
|
|||||||
"app_detail.html",
|
"app_detail.html",
|
||||||
"app_edit.html",
|
"app_edit.html",
|
||||||
"deployments.html",
|
"deployments.html",
|
||||||
"webhook_events.html",
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pageTemplates = make(map[string]*template.Template)
|
pageTemplates = make(map[string]*template.Template)
|
||||||
|
|||||||
@@ -1,34 +0,0 @@
|
|||||||
package templates //nolint:testpackage // tests unexported formatMemoryBytes
|
|
||||||
|
|
||||||
import (
|
|
||||||
"testing"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestFormatMemoryBytes(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
bytes int64
|
|
||||||
expected string
|
|
||||||
}{
|
|
||||||
{"gigabytes", 1024 * 1024 * 1024, "1g"},
|
|
||||||
{"two gigabytes", 2 * 1024 * 1024 * 1024, "2g"},
|
|
||||||
{"megabytes", 256 * 1024 * 1024, "256m"},
|
|
||||||
{"kilobytes", 512 * 1024, "512k"},
|
|
||||||
{"plain bytes", 12345, "12345"},
|
|
||||||
{"non-even megabytes", 256*1024*1024 + 1, "268435457"},
|
|
||||||
{"zero", 0, "0"},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
got := formatMemoryBytes(tt.bytes)
|
|
||||||
if got != tt.expected {
|
|
||||||
t.Errorf("formatMemoryBytes(%d) = %q, want %q", tt.bytes, got, tt.expected)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,79 +0,0 @@
|
|||||||
{{template "base" .}}
|
|
||||||
|
|
||||||
{{define "title"}}Webhook Events - {{.App.Name}} - µPaaS{{end}}
|
|
||||||
|
|
||||||
{{define "content"}}
|
|
||||||
{{template "nav" .}}
|
|
||||||
|
|
||||||
<main class="max-w-4xl mx-auto px-4 py-8">
|
|
||||||
<div class="mb-6">
|
|
||||||
<a href="/apps/{{.App.ID}}" class="text-primary-600 hover:text-primary-800 inline-flex items-center">
|
|
||||||
<svg class="w-4 h-4 mr-1" fill="none" stroke="currentColor" viewBox="0 0 24 24">
|
|
||||||
<path stroke-linecap="round" stroke-linejoin="round" stroke-width="2" d="M15 19l-7-7 7-7"/>
|
|
||||||
</svg>
|
|
||||||
Back to {{.App.Name}}
|
|
||||||
</a>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="section-header">
|
|
||||||
<h1 class="text-2xl font-medium text-gray-900">Webhook Events</h1>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
{{if .Events}}
|
|
||||||
<div class="card overflow-hidden">
|
|
||||||
<table class="table">
|
|
||||||
<thead class="table-header">
|
|
||||||
<tr>
|
|
||||||
<th>Time</th>
|
|
||||||
<th>Event</th>
|
|
||||||
<th>Branch</th>
|
|
||||||
<th>Commit</th>
|
|
||||||
<th>Status</th>
|
|
||||||
</tr>
|
|
||||||
</thead>
|
|
||||||
<tbody class="table-body">
|
|
||||||
{{range .Events}}
|
|
||||||
<tr>
|
|
||||||
<td class="text-gray-500 text-sm whitespace-nowrap">
|
|
||||||
<span x-data="relativeTime('{{.CreatedAt.Format `2006-01-02T15:04:05Z07:00`}}')" x-text="display" class="cursor-default" title="{{.CreatedAt.Format `2006-01-02 15:04:05`}}"></span>
|
|
||||||
</td>
|
|
||||||
<td class="text-gray-700 text-sm">{{.EventType}}</td>
|
|
||||||
<td class="font-mono text-gray-500 text-sm">{{.Branch}}</td>
|
|
||||||
<td class="font-mono text-gray-500 text-xs">
|
|
||||||
{{if and .CommitSHA.Valid .CommitURL.Valid}}
|
|
||||||
<a href="{{.CommitURL.String}}" target="_blank" rel="noopener noreferrer" class="text-primary-600 hover:text-primary-800">{{.ShortCommit}}</a>
|
|
||||||
{{else if .CommitSHA.Valid}}
|
|
||||||
{{.ShortCommit}}
|
|
||||||
{{else}}
|
|
||||||
<span class="text-gray-400">-</span>
|
|
||||||
{{end}}
|
|
||||||
</td>
|
|
||||||
<td>
|
|
||||||
{{if .Matched}}
|
|
||||||
{{if .Processed}}
|
|
||||||
<span class="badge-success">Matched</span>
|
|
||||||
{{else}}
|
|
||||||
<span class="badge-warning">Matched (pending)</span>
|
|
||||||
{{end}}
|
|
||||||
{{else}}
|
|
||||||
<span class="badge-neutral">No match</span>
|
|
||||||
{{end}}
|
|
||||||
</td>
|
|
||||||
</tr>
|
|
||||||
{{end}}
|
|
||||||
</tbody>
|
|
||||||
</table>
|
|
||||||
</div>
|
|
||||||
{{else}}
|
|
||||||
<div class="card">
|
|
||||||
<div class="empty-state">
|
|
||||||
<svg class="empty-state-icon" fill="none" stroke="currentColor" viewBox="0 0 24 24">
|
|
||||||
<path stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="M13 10V3L4 14h7v7l9-11h-7z"/>
|
|
||||||
</svg>
|
|
||||||
<h3 class="empty-state-title">No webhook events yet</h3>
|
|
||||||
<p class="empty-state-description">Webhook events will appear here once your repository sends push notifications.</p>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
{{end}}
|
|
||||||
</main>
|
|
||||||
{{end}}
|
|
||||||
Reference in New Issue
Block a user