Compare commits
1 Commits
fix/disabl
...
schema-con
| Author | SHA1 | Date | |
|---|---|---|---|
| 4e6542badf |
@@ -1,26 +0,0 @@
|
|||||||
name: Check
|
|
||||||
|
|
||||||
on:
|
|
||||||
push:
|
|
||||||
branches: [main]
|
|
||||||
pull_request:
|
|
||||||
branches: [main]
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
check:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@34e114876b0b11c390a56381ad16ebd13914f8d5 # v4
|
|
||||||
|
|
||||||
- uses: actions/setup-go@40f1582b2485089dde7abd97c1529aa768e1baff # v5
|
|
||||||
with:
|
|
||||||
go-version-file: go.mod
|
|
||||||
|
|
||||||
- name: Install golangci-lint
|
|
||||||
run: go install github.com/golangci/golangci-lint/v2/cmd/golangci-lint@5d1e709b7be35cb2025444e19de266b056b7b7ee # v2.10.1
|
|
||||||
|
|
||||||
- name: Install goimports
|
|
||||||
run: go install golang.org/x/tools/cmd/goimports@009367f5c17a8d4c45a961a3a509277190a9a6f0 # v0.42.0
|
|
||||||
|
|
||||||
- name: Run make check
|
|
||||||
run: make check
|
|
||||||
180
BUGS.md
Normal file
180
BUGS.md
Normal file
@@ -0,0 +1,180 @@
|
|||||||
|
# Bugs in µPaaS
|
||||||
|
|
||||||
|
## 1. Potential Race Condition in Log Writing
|
||||||
|
|
||||||
|
### Description
|
||||||
|
In the deployment service, when a deployment fails, the `failDeployment` function calls `writeLogsToFile` which may be called concurrently with the async log writer's flush operations. This could lead to partial or corrupted log files.
|
||||||
|
|
||||||
|
### Location
|
||||||
|
`internal/service/deploy/deploy.go:1169` in `failDeployment` function
|
||||||
|
|
||||||
|
### Proposed Fix
|
||||||
|
1. Add synchronization to ensure only one log write operation occurs at a time
|
||||||
|
2. Modify the `deploymentLogWriter` to track completion status and prevent concurrent writes
|
||||||
|
3. Add a wait mechanism in `failDeployment` to ensure any ongoing flush operations complete before writing logs to file
|
||||||
|
|
||||||
|
```go
|
||||||
|
// Add a mutex to deploymentLogWriter
|
||||||
|
type deploymentLogWriter struct {
|
||||||
|
// existing fields...
|
||||||
|
mu sync.Mutex
|
||||||
|
writeMu sync.Mutex // Add this for file writing synchronization
|
||||||
|
done chan struct{}
|
||||||
|
flushed sync.WaitGroup
|
||||||
|
}
|
||||||
|
|
||||||
|
// In writeLogsToFile, ensure exclusive access
|
||||||
|
func (svc *Service) writeLogsToFile(app *models.App, deployment *models.Deployment) {
|
||||||
|
svc.writeMu.Lock() // Add this mutex to Service struct
|
||||||
|
defer svc.writeMu.Unlock()
|
||||||
|
// existing code...
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
## 2. Incomplete Error Handling in Container Operations
|
||||||
|
|
||||||
|
### Description
|
||||||
|
In the Docker client's `performClone` function, if `createGitContainer` fails, the SSH key file created earlier is not cleaned up, causing a potential security risk.
|
||||||
|
|
||||||
|
### Location
|
||||||
|
`internal/docker/client.go:597` in `performClone` function
|
||||||
|
|
||||||
|
### Proposed Fix
|
||||||
|
Add proper cleanup using `defer` immediately after creating the SSH key file:
|
||||||
|
|
||||||
|
```go
|
||||||
|
// After writing SSH key file (line 578)
|
||||||
|
keyFileCreated := false
|
||||||
|
err = os.WriteFile(cfg.keyFile, []byte(cfg.sshPrivateKey), sshKeyPermissions)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to write SSH key: %w", err)
|
||||||
|
}
|
||||||
|
keyFileCreated = true
|
||||||
|
|
||||||
|
defer func() {
|
||||||
|
if keyFileCreated {
|
||||||
|
removeErr := os.Remove(cfg.keyFile)
|
||||||
|
if removeErr != nil {
|
||||||
|
c.log.Error("failed to remove SSH key file", "error", removeErr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
```
|
||||||
|
|
||||||
|
## 3. Missing Context Cancellation Check During Build
|
||||||
|
|
||||||
|
### Description
|
||||||
|
In the deployment service's `streamBuildOutput` function, long-running Docker build operations may not properly respond to context cancellation, causing deployments to hang even when cancelled.
|
||||||
|
|
||||||
|
### Location
|
||||||
|
`internal/docker/client.go:542` in `streamBuildOutput` function
|
||||||
|
|
||||||
|
### Proposed Fix
|
||||||
|
Add context checking in the scanner loop:
|
||||||
|
|
||||||
|
```go
|
||||||
|
for scanner.Scan() {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return ctx.Err()
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
|
||||||
|
line := scanner.Bytes()
|
||||||
|
// existing code...
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
## 4. Inconsistent Container Removal in Error Cases
|
||||||
|
|
||||||
|
### Description
|
||||||
|
When deployment fails during container creation, the already-created container is not removed, leading to orphaned containers that consume resources.
|
||||||
|
|
||||||
|
### Location
|
||||||
|
`internal/service/deploy/deploy.go:969` in `createAndStartContainer` function
|
||||||
|
|
||||||
|
### Proposed Fix
|
||||||
|
Add cleanup of created container on start failure:
|
||||||
|
|
||||||
|
```go
|
||||||
|
containerID, err := svc.docker.CreateContainer(ctx, containerOpts)
|
||||||
|
if err != nil {
|
||||||
|
svc.notify.NotifyDeployFailed(ctx, app, deployment, err)
|
||||||
|
svc.failDeployment(ctx, app, deployment, fmt.Errorf("failed to create container: %w", err))
|
||||||
|
return "", fmt.Errorf("failed to create container: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add cleanup defer for error cases
|
||||||
|
defer func() {
|
||||||
|
if err != nil {
|
||||||
|
// If we have a container ID but returning an error, clean it up
|
||||||
|
_ = svc.docker.RemoveContainer(context.Background(), containerID, true)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
startErr := svc.docker.StartContainer(ctx, containerID)
|
||||||
|
if startErr != nil {
|
||||||
|
svc.notify.NotifyDeployFailed(ctx, app, deployment, startErr)
|
||||||
|
svc.failDeployment(ctx, app, deployment, fmt.Errorf("failed to start container: %w", startErr))
|
||||||
|
err = startErr // Set err so defer cleanup runs
|
||||||
|
return "", fmt.Errorf("failed to start container: %w", startErr)
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
## 5. Potential Data Race in Active Deployments Tracking
|
||||||
|
|
||||||
|
### Description
|
||||||
|
The `activeDeploys` sync.Map in the deployment service may have race conditions when multiple concurrent deployments try to access the same app's deployment state.
|
||||||
|
|
||||||
|
### Location
|
||||||
|
`internal/service/deploy/deploy.go:226` and related functions
|
||||||
|
|
||||||
|
### Proposed Fix
|
||||||
|
Add proper locking around active deploy operations:
|
||||||
|
|
||||||
|
```go
|
||||||
|
// Add a mutex for active deploy operations
|
||||||
|
type Service struct {
|
||||||
|
// existing fields...
|
||||||
|
activeDeployMu sync.Mutex
|
||||||
|
}
|
||||||
|
|
||||||
|
// In Deploy function
|
||||||
|
func (svc *Service) Deploy(ctx context.Context, app *models.App, webhookEventID *int64, cancelExisting bool) error {
|
||||||
|
svc.activeDeployMu.Lock()
|
||||||
|
if cancelExisting {
|
||||||
|
svc.cancelActiveDeploy(app.ID)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Try to acquire per-app deployment lock
|
||||||
|
if !svc.tryLockApp(app.ID) {
|
||||||
|
svc.activeDeployMu.Unlock()
|
||||||
|
svc.log.Warn("deployment already in progress", "app", app.Name)
|
||||||
|
return ErrDeploymentInProgress
|
||||||
|
}
|
||||||
|
svc.activeDeployMu.Unlock()
|
||||||
|
|
||||||
|
defer svc.unlockApp(app.ID)
|
||||||
|
// rest of function...
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
## 6. Incomplete Error Propagation in Git Clone
|
||||||
|
|
||||||
|
### Description
|
||||||
|
In the Docker client's `runGitClone` function, if `ContainerLogs` fails, the error is silently ignored, which could hide important debugging information.
|
||||||
|
|
||||||
|
### Location
|
||||||
|
`internal/docker/client.go:679` in `runGitClone` function
|
||||||
|
|
||||||
|
### Proposed Fix
|
||||||
|
Handle the ContainerLogs error properly:
|
||||||
|
|
||||||
|
```go
|
||||||
|
// Always capture logs for the result
|
||||||
|
logs, logErr := c.ContainerLogs(ctx, containerID, "100")
|
||||||
|
if logErr != nil {
|
||||||
|
c.log.Warn("failed to get git clone logs", "error", logErr)
|
||||||
|
logs = "Failed to retrieve logs: " + logErr.Error()
|
||||||
|
}
|
||||||
|
```
|
||||||
68
CLAUDE.md
Normal file
68
CLAUDE.md
Normal file
@@ -0,0 +1,68 @@
|
|||||||
|
# Repository Rules
|
||||||
|
|
||||||
|
Last Updated 2026-01-08
|
||||||
|
|
||||||
|
These rules MUST be followed at all times, it is very important.
|
||||||
|
|
||||||
|
* Never use `git add -A` - add specific changes to a deliberate commit. A
|
||||||
|
commit should contain one change. After each change, make a commit with a
|
||||||
|
good one-line summary.
|
||||||
|
|
||||||
|
* NEVER modify the linter config without asking first.
|
||||||
|
|
||||||
|
* NEVER modify tests to exclude special cases or otherwise get them to pass
|
||||||
|
without asking first. In almost all cases, the code should be changed,
|
||||||
|
NOT the tests. If you think the test needs to be changed, make your case
|
||||||
|
for that and ask for permission to proceed, then stop. You need explicit
|
||||||
|
user approval to modify existing tests. (You do not need user approval
|
||||||
|
for writing NEW tests.)
|
||||||
|
|
||||||
|
* When linting, assume the linter config is CORRECT, and that each item
|
||||||
|
output by the linter is something that legitimately needs fixing in the
|
||||||
|
code.
|
||||||
|
|
||||||
|
* When running tests, use `make test`.
|
||||||
|
|
||||||
|
* Before commits, run `make check`. This runs `make lint` and `make test`
|
||||||
|
and `make check-fmt`. Any issues discovered MUST be resolved before
|
||||||
|
committing unless explicitly told otherwise.
|
||||||
|
|
||||||
|
* When fixing a bug, write a failing test for the bug FIRST. Add
|
||||||
|
appropriate logging to the test to ensure it is written correctly. Commit
|
||||||
|
that. Then go about fixing the bug until the test passes (without
|
||||||
|
modifying the test further). Then commit that.
|
||||||
|
|
||||||
|
* When adding a new feature, do the same - implement a test first (TDD). It
|
||||||
|
doesn't have to be super complex. Commit the test, then commit the
|
||||||
|
feature.
|
||||||
|
|
||||||
|
* When adding a new feature, use a feature branch. When the feature is
|
||||||
|
completely finished and the code is up to standards (passes `make check`)
|
||||||
|
then and only then can the feature branch be merged into `main` and the
|
||||||
|
branch deleted.
|
||||||
|
|
||||||
|
* Write godoc documentation comments for all exported types and functions as
|
||||||
|
you go along.
|
||||||
|
|
||||||
|
* ALWAYS be consistent in naming. If you name something one thing in one
|
||||||
|
place, name it the EXACT SAME THING in another place.
|
||||||
|
|
||||||
|
* Be descriptive and specific in naming. `wl` is bad;
|
||||||
|
`SourceHostWhitelist` is good. `ConnsPerHost` is bad;
|
||||||
|
`MaxConnectionsPerHost` is good.
|
||||||
|
|
||||||
|
* This is not prototype or teaching code - this is designed for production.
|
||||||
|
Any security issues (such as denial of service) or other web
|
||||||
|
vulnerabilities are P1 bugs and must be added to TODO.md at the top.
|
||||||
|
|
||||||
|
* As this is production code, no stubbing of implementations unless
|
||||||
|
specifically instructed. We need working implementations.
|
||||||
|
|
||||||
|
* Avoid vendoring deps unless specifically instructed to. NEVER commit
|
||||||
|
the vendor directory, NEVER commit compiled binaries. If these
|
||||||
|
directories or files exist, add them to .gitignore (and commit the
|
||||||
|
.gitignore) if they are not already in there. Keep the entire git
|
||||||
|
repository (with history) small - under 20MiB, unless you specifically
|
||||||
|
must commit larger files (e.g. test fixture example media files). Only
|
||||||
|
OUR source code and immediately supporting files (such as test examples)
|
||||||
|
goes into the repo/history.
|
||||||
@@ -51,8 +51,7 @@ type Config struct {
|
|||||||
MaintenanceMode bool
|
MaintenanceMode bool
|
||||||
MetricsUsername string
|
MetricsUsername string
|
||||||
MetricsPassword string
|
MetricsPassword string
|
||||||
SessionSecret string `json:"-"`
|
SessionSecret string
|
||||||
CORSOrigins string
|
|
||||||
params *Params
|
params *Params
|
||||||
log *slog.Logger
|
log *slog.Logger
|
||||||
}
|
}
|
||||||
@@ -103,7 +102,6 @@ func setupViper(name string) {
|
|||||||
viper.SetDefault("METRICS_USERNAME", "")
|
viper.SetDefault("METRICS_USERNAME", "")
|
||||||
viper.SetDefault("METRICS_PASSWORD", "")
|
viper.SetDefault("METRICS_PASSWORD", "")
|
||||||
viper.SetDefault("SESSION_SECRET", "")
|
viper.SetDefault("SESSION_SECRET", "")
|
||||||
viper.SetDefault("CORS_ORIGINS", "")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func buildConfig(log *slog.Logger, params *Params) (*Config, error) {
|
func buildConfig(log *slog.Logger, params *Params) (*Config, error) {
|
||||||
@@ -138,7 +136,6 @@ func buildConfig(log *slog.Logger, params *Params) (*Config, error) {
|
|||||||
MetricsUsername: viper.GetString("METRICS_USERNAME"),
|
MetricsUsername: viper.GetString("METRICS_USERNAME"),
|
||||||
MetricsPassword: viper.GetString("METRICS_PASSWORD"),
|
MetricsPassword: viper.GetString("METRICS_PASSWORD"),
|
||||||
SessionSecret: viper.GetString("SESSION_SECRET"),
|
SessionSecret: viper.GetString("SESSION_SECRET"),
|
||||||
CORSOrigins: viper.GetString("CORS_ORIGINS"),
|
|
||||||
params: params,
|
params: params,
|
||||||
log: log,
|
log: log,
|
||||||
}
|
}
|
||||||
|
|||||||
6
internal/database/migrations/001_init_migrations.sql
Normal file
6
internal/database/migrations/001_init_migrations.sql
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
-- Initialize migrations table for tracking applied migrations
|
||||||
|
CREATE TABLE IF NOT EXISTS migrations (
|
||||||
|
id INTEGER PRIMARY KEY,
|
||||||
|
name TEXT NOT NULL UNIQUE,
|
||||||
|
applied_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
@@ -1,7 +1,8 @@
|
|||||||
-- Initial schema for upaas
|
-- Complete schema for upaas (consolidated)
|
||||||
|
-- This represents the final state of all migrations applied
|
||||||
|
|
||||||
-- Users table (single admin user)
|
-- Users table (single admin user)
|
||||||
CREATE TABLE users (
|
CREATE TABLE IF NOT EXISTS users (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
username TEXT UNIQUE NOT NULL,
|
username TEXT UNIQUE NOT NULL,
|
||||||
password_hash TEXT NOT NULL,
|
password_hash TEXT NOT NULL,
|
||||||
@@ -9,7 +10,7 @@ CREATE TABLE users (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- Apps table
|
-- Apps table
|
||||||
CREATE TABLE apps (
|
CREATE TABLE IF NOT EXISTS apps (
|
||||||
id TEXT PRIMARY KEY,
|
id TEXT PRIMARY KEY,
|
||||||
name TEXT UNIQUE NOT NULL,
|
name TEXT UNIQUE NOT NULL,
|
||||||
repo_url TEXT NOT NULL,
|
repo_url TEXT NOT NULL,
|
||||||
@@ -18,18 +19,19 @@ CREATE TABLE apps (
|
|||||||
webhook_secret TEXT NOT NULL,
|
webhook_secret TEXT NOT NULL,
|
||||||
ssh_private_key TEXT NOT NULL,
|
ssh_private_key TEXT NOT NULL,
|
||||||
ssh_public_key TEXT NOT NULL,
|
ssh_public_key TEXT NOT NULL,
|
||||||
container_id TEXT,
|
|
||||||
image_id TEXT,
|
image_id TEXT,
|
||||||
|
previous_image_id TEXT,
|
||||||
status TEXT DEFAULT 'pending',
|
status TEXT DEFAULT 'pending',
|
||||||
docker_network TEXT,
|
docker_network TEXT,
|
||||||
ntfy_topic TEXT,
|
ntfy_topic TEXT,
|
||||||
slack_webhook TEXT,
|
slack_webhook TEXT,
|
||||||
|
webhook_secret_hash TEXT NOT NULL DEFAULT '',
|
||||||
created_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
created_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
||||||
updated_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
updated_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
||||||
);
|
);
|
||||||
|
|
||||||
-- App environment variables
|
-- App environment variables
|
||||||
CREATE TABLE app_env_vars (
|
CREATE TABLE IF NOT EXISTS app_env_vars (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
key TEXT NOT NULL,
|
key TEXT NOT NULL,
|
||||||
@@ -38,7 +40,7 @@ CREATE TABLE app_env_vars (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- App labels
|
-- App labels
|
||||||
CREATE TABLE app_labels (
|
CREATE TABLE IF NOT EXISTS app_labels (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
key TEXT NOT NULL,
|
key TEXT NOT NULL,
|
||||||
@@ -47,7 +49,7 @@ CREATE TABLE app_labels (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- App volume mounts
|
-- App volume mounts
|
||||||
CREATE TABLE app_volumes (
|
CREATE TABLE IF NOT EXISTS app_volumes (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
host_path TEXT NOT NULL,
|
host_path TEXT NOT NULL,
|
||||||
@@ -55,13 +57,24 @@ CREATE TABLE app_volumes (
|
|||||||
readonly INTEGER DEFAULT 0
|
readonly INTEGER DEFAULT 0
|
||||||
);
|
);
|
||||||
|
|
||||||
|
-- App port mappings
|
||||||
|
CREATE TABLE IF NOT EXISTS app_ports (
|
||||||
|
id INTEGER PRIMARY KEY,
|
||||||
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
|
host_port INTEGER NOT NULL,
|
||||||
|
container_port INTEGER NOT NULL,
|
||||||
|
protocol TEXT NOT NULL DEFAULT 'tcp' CHECK(protocol IN ('tcp', 'udp')),
|
||||||
|
UNIQUE(host_port, protocol)
|
||||||
|
);
|
||||||
|
|
||||||
-- Webhook events log
|
-- Webhook events log
|
||||||
CREATE TABLE webhook_events (
|
CREATE TABLE IF NOT EXISTS webhook_events (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
event_type TEXT NOT NULL,
|
event_type TEXT NOT NULL,
|
||||||
branch TEXT NOT NULL,
|
branch TEXT NOT NULL,
|
||||||
commit_sha TEXT,
|
commit_sha TEXT,
|
||||||
|
commit_url TEXT,
|
||||||
payload TEXT,
|
payload TEXT,
|
||||||
matched INTEGER NOT NULL,
|
matched INTEGER NOT NULL,
|
||||||
processed INTEGER DEFAULT 0,
|
processed INTEGER DEFAULT 0,
|
||||||
@@ -69,13 +82,13 @@ CREATE TABLE webhook_events (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- Deployments log
|
-- Deployments log
|
||||||
CREATE TABLE deployments (
|
CREATE TABLE IF NOT EXISTS deployments (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
webhook_event_id INTEGER REFERENCES webhook_events(id),
|
webhook_event_id INTEGER REFERENCES webhook_events(id),
|
||||||
commit_sha TEXT,
|
commit_sha TEXT,
|
||||||
|
commit_url TEXT,
|
||||||
image_id TEXT,
|
image_id TEXT,
|
||||||
container_id TEXT,
|
|
||||||
status TEXT NOT NULL,
|
status TEXT NOT NULL,
|
||||||
logs TEXT,
|
logs TEXT,
|
||||||
started_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
started_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
||||||
@@ -83,12 +96,14 @@ CREATE TABLE deployments (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- Indexes
|
-- Indexes
|
||||||
CREATE INDEX idx_apps_status ON apps(status);
|
CREATE INDEX IF NOT EXISTS idx_apps_status ON apps(status);
|
||||||
CREATE INDEX idx_apps_webhook_secret ON apps(webhook_secret);
|
CREATE INDEX IF NOT EXISTS idx_apps_webhook_secret ON apps(webhook_secret);
|
||||||
CREATE INDEX idx_app_env_vars_app_id ON app_env_vars(app_id);
|
CREATE INDEX IF NOT EXISTS idx_apps_webhook_secret_hash ON apps(webhook_secret_hash);
|
||||||
CREATE INDEX idx_app_labels_app_id ON app_labels(app_id);
|
CREATE INDEX IF NOT EXISTS idx_app_env_vars_app_id ON app_env_vars(app_id);
|
||||||
CREATE INDEX idx_app_volumes_app_id ON app_volumes(app_id);
|
CREATE INDEX IF NOT EXISTS idx_app_labels_app_id ON app_labels(app_id);
|
||||||
CREATE INDEX idx_webhook_events_app_id ON webhook_events(app_id);
|
CREATE INDEX IF NOT EXISTS idx_app_volumes_app_id ON app_volumes(app_id);
|
||||||
CREATE INDEX idx_webhook_events_created_at ON webhook_events(created_at);
|
CREATE INDEX IF NOT EXISTS idx_app_ports_app_id ON app_ports(app_id);
|
||||||
CREATE INDEX idx_deployments_app_id ON deployments(app_id);
|
CREATE INDEX IF NOT EXISTS idx_webhook_events_app_id ON webhook_events(app_id);
|
||||||
CREATE INDEX idx_deployments_started_at ON deployments(started_at);
|
CREATE INDEX IF NOT EXISTS idx_webhook_events_created_at ON webhook_events(created_at);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_deployments_app_id ON deployments(app_id);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_deployments_started_at ON deployments(started_at);
|
||||||
@@ -1,44 +0,0 @@
|
|||||||
-- Remove container_id from apps table
|
|
||||||
-- Container is now looked up via Docker label (upaas.id) instead of stored in database
|
|
||||||
|
|
||||||
-- SQLite doesn't support DROP COLUMN before version 3.35.0 (2021-03-12)
|
|
||||||
-- Use table rebuild for broader compatibility
|
|
||||||
|
|
||||||
-- Create new table without container_id
|
|
||||||
CREATE TABLE apps_new (
|
|
||||||
id TEXT PRIMARY KEY,
|
|
||||||
name TEXT UNIQUE NOT NULL,
|
|
||||||
repo_url TEXT NOT NULL,
|
|
||||||
branch TEXT NOT NULL DEFAULT 'main',
|
|
||||||
dockerfile_path TEXT DEFAULT 'Dockerfile',
|
|
||||||
webhook_secret TEXT NOT NULL,
|
|
||||||
ssh_private_key TEXT NOT NULL,
|
|
||||||
ssh_public_key TEXT NOT NULL,
|
|
||||||
image_id TEXT,
|
|
||||||
status TEXT DEFAULT 'pending',
|
|
||||||
docker_network TEXT,
|
|
||||||
ntfy_topic TEXT,
|
|
||||||
slack_webhook TEXT,
|
|
||||||
created_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
|
||||||
updated_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
|
||||||
);
|
|
||||||
|
|
||||||
-- Copy data (excluding container_id)
|
|
||||||
INSERT INTO apps_new (
|
|
||||||
id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
|
||||||
ssh_private_key, ssh_public_key, image_id, status,
|
|
||||||
docker_network, ntfy_topic, slack_webhook, created_at, updated_at
|
|
||||||
)
|
|
||||||
SELECT
|
|
||||||
id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
|
||||||
ssh_private_key, ssh_public_key, image_id, status,
|
|
||||||
docker_network, ntfy_topic, slack_webhook, created_at, updated_at
|
|
||||||
FROM apps;
|
|
||||||
|
|
||||||
-- Drop old table and rename new one
|
|
||||||
DROP TABLE apps;
|
|
||||||
ALTER TABLE apps_new RENAME TO apps;
|
|
||||||
|
|
||||||
-- Recreate indexes
|
|
||||||
CREATE INDEX idx_apps_status ON apps(status);
|
|
||||||
CREATE INDEX idx_apps_webhook_secret ON apps(webhook_secret);
|
|
||||||
@@ -1,12 +0,0 @@
|
|||||||
-- Add port mappings for apps
|
|
||||||
|
|
||||||
CREATE TABLE app_ports (
|
|
||||||
id INTEGER PRIMARY KEY,
|
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
|
||||||
host_port INTEGER NOT NULL,
|
|
||||||
container_port INTEGER NOT NULL,
|
|
||||||
protocol TEXT NOT NULL DEFAULT 'tcp' CHECK(protocol IN ('tcp', 'udp')),
|
|
||||||
UNIQUE(host_port, protocol)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX idx_app_ports_app_id ON app_ports(app_id);
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
-- Add commit_url column to webhook_events and deployments tables
|
|
||||||
ALTER TABLE webhook_events ADD COLUMN commit_url TEXT;
|
|
||||||
ALTER TABLE deployments ADD COLUMN commit_url TEXT;
|
|
||||||
@@ -1,2 +0,0 @@
|
|||||||
-- Add webhook_secret_hash column for constant-time secret lookup
|
|
||||||
ALTER TABLE apps ADD COLUMN webhook_secret_hash TEXT NOT NULL DEFAULT '';
|
|
||||||
@@ -1,2 +0,0 @@
|
|||||||
-- Add previous_image_id to apps for deployment rollback support
|
|
||||||
ALTER TABLE apps ADD COLUMN previous_image_id TEXT;
|
|
||||||
@@ -17,7 +17,6 @@ import (
|
|||||||
"github.com/docker/docker/api/types"
|
"github.com/docker/docker/api/types"
|
||||||
"github.com/docker/docker/api/types/container"
|
"github.com/docker/docker/api/types/container"
|
||||||
"github.com/docker/docker/api/types/filters"
|
"github.com/docker/docker/api/types/filters"
|
||||||
"github.com/docker/docker/api/types/image"
|
|
||||||
"github.com/docker/docker/api/types/mount"
|
"github.com/docker/docker/api/types/mount"
|
||||||
"github.com/docker/docker/api/types/network"
|
"github.com/docker/docker/api/types/network"
|
||||||
"github.com/docker/docker/client"
|
"github.com/docker/docker/client"
|
||||||
@@ -480,20 +479,6 @@ func (c *Client) CloneRepo(
|
|||||||
return c.performClone(ctx, cfg)
|
return c.performClone(ctx, cfg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// RemoveImage removes a Docker image by ID or tag.
|
|
||||||
// It returns nil if the image was successfully removed or does not exist.
|
|
||||||
func (c *Client) RemoveImage(ctx context.Context, imageID string) error {
|
|
||||||
_, err := c.docker.ImageRemove(ctx, imageID, image.RemoveOptions{
|
|
||||||
Force: true,
|
|
||||||
PruneChildren: true,
|
|
||||||
})
|
|
||||||
if err != nil && !client.IsErrNotFound(err) {
|
|
||||||
return fmt.Errorf("failed to remove image %s: %w", imageID, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *Client) performBuild(
|
func (c *Client) performBuild(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
opts BuildImageOptions,
|
opts BuildImageOptions,
|
||||||
|
|||||||
@@ -8,6 +8,7 @@ import (
|
|||||||
"github.com/go-chi/chi/v5"
|
"github.com/go-chi/chi/v5"
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/service/app"
|
||||||
)
|
)
|
||||||
|
|
||||||
// apiAppResponse is the JSON representation of an app.
|
// apiAppResponse is the JSON representation of an app.
|
||||||
@@ -73,13 +74,18 @@ func deploymentToAPI(d *models.Deployment) apiDeploymentResponse {
|
|||||||
// HandleAPILoginPOST returns a handler that authenticates via JSON credentials
|
// HandleAPILoginPOST returns a handler that authenticates via JSON credentials
|
||||||
// and sets a session cookie.
|
// and sets a session cookie.
|
||||||
func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
||||||
|
type loginRequest struct {
|
||||||
|
Username string `json:"username"`
|
||||||
|
Password string `json:"password"`
|
||||||
|
}
|
||||||
|
|
||||||
type loginResponse struct {
|
type loginResponse struct {
|
||||||
UserID int64 `json:"userId"`
|
UserID int64 `json:"userId"`
|
||||||
Username string `json:"username"`
|
Username string `json:"username"`
|
||||||
}
|
}
|
||||||
|
|
||||||
return func(writer http.ResponseWriter, request *http.Request) {
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
var req map[string]string
|
var req loginRequest
|
||||||
|
|
||||||
decodeErr := json.NewDecoder(request.Body).Decode(&req)
|
decodeErr := json.NewDecoder(request.Body).Decode(&req)
|
||||||
if decodeErr != nil {
|
if decodeErr != nil {
|
||||||
@@ -90,10 +96,7 @@ func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
username := req["username"]
|
if req.Username == "" || req.Password == "" {
|
||||||
credential := req["password"]
|
|
||||||
|
|
||||||
if username == "" || credential == "" {
|
|
||||||
h.respondJSON(writer, request,
|
h.respondJSON(writer, request,
|
||||||
map[string]string{"error": "username and password are required"},
|
map[string]string{"error": "username and password are required"},
|
||||||
http.StatusBadRequest)
|
http.StatusBadRequest)
|
||||||
@@ -101,7 +104,7 @@ func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
user, authErr := h.auth.Authenticate(request.Context(), username, credential)
|
user, authErr := h.auth.Authenticate(request.Context(), req.Username, req.Password)
|
||||||
if authErr != nil {
|
if authErr != nil {
|
||||||
h.respondJSON(writer, request,
|
h.respondJSON(writer, request,
|
||||||
map[string]string{"error": "invalid credentials"},
|
map[string]string{"error": "invalid credentials"},
|
||||||
@@ -174,6 +177,106 @@ func (h *Handlers) HandleAPIGetApp() http.HandlerFunc {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// HandleAPICreateApp returns a handler that creates a new app.
|
||||||
|
func (h *Handlers) HandleAPICreateApp() http.HandlerFunc {
|
||||||
|
type createRequest struct {
|
||||||
|
Name string `json:"name"`
|
||||||
|
RepoURL string `json:"repoUrl"`
|
||||||
|
Branch string `json:"branch"`
|
||||||
|
DockerfilePath string `json:"dockerfilePath"`
|
||||||
|
DockerNetwork string `json:"dockerNetwork"`
|
||||||
|
NtfyTopic string `json:"ntfyTopic"`
|
||||||
|
SlackWebhook string `json:"slackWebhook"`
|
||||||
|
}
|
||||||
|
|
||||||
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
var req createRequest
|
||||||
|
|
||||||
|
decodeErr := json.NewDecoder(request.Body).Decode(&req)
|
||||||
|
if decodeErr != nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "invalid JSON body"},
|
||||||
|
http.StatusBadRequest)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if req.Name == "" || req.RepoURL == "" {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "name and repo_url are required"},
|
||||||
|
http.StatusBadRequest)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
nameErr := validateAppName(req.Name)
|
||||||
|
if nameErr != nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "invalid app name: " + nameErr.Error()},
|
||||||
|
http.StatusBadRequest)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
createdApp, createErr := h.appService.CreateApp(request.Context(), app.CreateAppInput{
|
||||||
|
Name: req.Name,
|
||||||
|
RepoURL: req.RepoURL,
|
||||||
|
Branch: req.Branch,
|
||||||
|
DockerfilePath: req.DockerfilePath,
|
||||||
|
DockerNetwork: req.DockerNetwork,
|
||||||
|
NtfyTopic: req.NtfyTopic,
|
||||||
|
SlackWebhook: req.SlackWebhook,
|
||||||
|
})
|
||||||
|
if createErr != nil {
|
||||||
|
h.log.Error("api: failed to create app", "error", createErr)
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "failed to create app"},
|
||||||
|
http.StatusInternalServerError)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
h.respondJSON(writer, request, appToAPI(createdApp), http.StatusCreated)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// HandleAPIDeleteApp returns a handler that deletes an app.
|
||||||
|
func (h *Handlers) HandleAPIDeleteApp() http.HandlerFunc {
|
||||||
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
appID := chi.URLParam(request, "id")
|
||||||
|
|
||||||
|
application, err := h.appService.GetApp(request.Context(), appID)
|
||||||
|
if err != nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "internal server error"},
|
||||||
|
http.StatusInternalServerError)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if application == nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "app not found"},
|
||||||
|
http.StatusNotFound)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
deleteErr := h.appService.DeleteApp(request.Context(), application)
|
||||||
|
if deleteErr != nil {
|
||||||
|
h.log.Error("api: failed to delete app", "error", deleteErr)
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "failed to delete app"},
|
||||||
|
http.StatusInternalServerError)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"status": "deleted"}, http.StatusOK)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// deploymentsPageLimit is the default number of deployments per page.
|
// deploymentsPageLimit is the default number of deployments per page.
|
||||||
const deploymentsPageLimit = 20
|
const deploymentsPageLimit = 20
|
||||||
|
|
||||||
@@ -220,6 +323,35 @@ func (h *Handlers) HandleAPIListDeployments() http.HandlerFunc {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// HandleAPITriggerDeploy returns a handler that triggers a deployment for an app.
|
||||||
|
func (h *Handlers) HandleAPITriggerDeploy() http.HandlerFunc {
|
||||||
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
appID := chi.URLParam(request, "id")
|
||||||
|
|
||||||
|
application, err := h.appService.GetApp(request.Context(), appID)
|
||||||
|
if err != nil || application == nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "app not found"},
|
||||||
|
http.StatusNotFound)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
deployErr := h.deploy.Deploy(request.Context(), application, nil, true)
|
||||||
|
if deployErr != nil {
|
||||||
|
h.log.Error("api: failed to trigger deploy", "error", deployErr)
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": deployErr.Error()},
|
||||||
|
http.StatusConflict)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"status": "deploying"}, http.StatusAccepted)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// HandleAPIWhoAmI returns a handler that shows the current authenticated user.
|
// HandleAPIWhoAmI returns a handler that shows the current authenticated user.
|
||||||
func (h *Handlers) HandleAPIWhoAmI() http.HandlerFunc {
|
func (h *Handlers) HandleAPIWhoAmI() http.HandlerFunc {
|
||||||
type whoAmIResponse struct {
|
type whoAmIResponse struct {
|
||||||
|
|||||||
@@ -10,8 +10,6 @@ import (
|
|||||||
"github.com/go-chi/chi/v5"
|
"github.com/go-chi/chi/v5"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/service/app"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// apiRouter builds a chi router with the API routes using session auth middleware.
|
// apiRouter builds a chi router with the API routes using session auth middleware.
|
||||||
@@ -25,7 +23,10 @@ func apiRouter(tc *testContext) http.Handler {
|
|||||||
apiR.Use(tc.middleware.APISessionAuth())
|
apiR.Use(tc.middleware.APISessionAuth())
|
||||||
apiR.Get("/whoami", tc.handlers.HandleAPIWhoAmI())
|
apiR.Get("/whoami", tc.handlers.HandleAPIWhoAmI())
|
||||||
apiR.Get("/apps", tc.handlers.HandleAPIListApps())
|
apiR.Get("/apps", tc.handlers.HandleAPIListApps())
|
||||||
|
apiR.Post("/apps", tc.handlers.HandleAPICreateApp())
|
||||||
apiR.Get("/apps/{id}", tc.handlers.HandleAPIGetApp())
|
apiR.Get("/apps/{id}", tc.handlers.HandleAPIGetApp())
|
||||||
|
apiR.Delete("/apps/{id}", tc.handlers.HandleAPIDeleteApp())
|
||||||
|
apiR.Post("/apps/{id}/deploy", tc.handlers.HandleAPITriggerDeploy())
|
||||||
apiR.Get("/apps/{id}/deployments", tc.handlers.HandleAPIListDeployments())
|
apiR.Get("/apps/{id}/deployments", tc.handlers.HandleAPIListDeployments())
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
@@ -61,16 +62,23 @@ func setupAPITest(t *testing.T) (*testContext, []*http.Cookie) {
|
|||||||
return tc, cookies
|
return tc, cookies
|
||||||
}
|
}
|
||||||
|
|
||||||
// apiGet makes an authenticated GET request using session cookies.
|
// apiRequest makes an authenticated API request using session cookies.
|
||||||
func apiGet(
|
func apiRequest(
|
||||||
t *testing.T,
|
t *testing.T,
|
||||||
tc *testContext,
|
tc *testContext,
|
||||||
cookies []*http.Cookie,
|
cookies []*http.Cookie,
|
||||||
path string,
|
method, path string,
|
||||||
|
body string,
|
||||||
) *httptest.ResponseRecorder {
|
) *httptest.ResponseRecorder {
|
||||||
t.Helper()
|
t.Helper()
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, path, nil)
|
var req *http.Request
|
||||||
|
if body != "" {
|
||||||
|
req = httptest.NewRequest(method, path, strings.NewReader(body))
|
||||||
|
req.Header.Set("Content-Type", "application/json")
|
||||||
|
} else {
|
||||||
|
req = httptest.NewRequest(method, path, nil)
|
||||||
|
}
|
||||||
|
|
||||||
for _, c := range cookies {
|
for _, c := range cookies {
|
||||||
req.AddCookie(c)
|
req.AddCookie(c)
|
||||||
@@ -167,7 +175,7 @@ func TestAPIWhoAmI(t *testing.T) {
|
|||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/whoami")
|
rr := apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/whoami", "")
|
||||||
assert.Equal(t, http.StatusOK, rr.Code)
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
var resp map[string]any
|
var resp map[string]any
|
||||||
@@ -180,7 +188,7 @@ func TestAPIListAppsEmpty(t *testing.T) {
|
|||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/apps")
|
rr := apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps", "")
|
||||||
assert.Equal(t, http.StatusOK, rr.Code)
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
var apps []any
|
var apps []any
|
||||||
@@ -188,23 +196,52 @@ func TestAPIListAppsEmpty(t *testing.T) {
|
|||||||
assert.Empty(t, apps)
|
assert.Empty(t, apps)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestAPICreateApp(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
|
body := `{"name":"test-app","repoUrl":"https://github.com/example/repo"}`
|
||||||
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
|
assert.Equal(t, http.StatusCreated, rr.Code)
|
||||||
|
|
||||||
|
var app map[string]any
|
||||||
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &app))
|
||||||
|
assert.Equal(t, "test-app", app["name"])
|
||||||
|
assert.Equal(t, "pending", app["status"])
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestAPICreateAppValidation(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
|
body := `{"name":"","repoUrl":""}`
|
||||||
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
|
assert.Equal(t, http.StatusBadRequest, rr.Code)
|
||||||
|
}
|
||||||
|
|
||||||
func TestAPIGetApp(t *testing.T) {
|
func TestAPIGetApp(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
created, err := tc.appSvc.CreateApp(t.Context(), app.CreateAppInput{
|
body := `{"name":"my-app","repoUrl":"https://github.com/example/repo"}`
|
||||||
Name: "my-app",
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
RepoURL: "https://github.com/example/repo",
|
require.Equal(t, http.StatusCreated, rr.Code)
|
||||||
})
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/apps/"+created.ID)
|
var created map[string]any
|
||||||
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &created))
|
||||||
|
|
||||||
|
appID, ok := created["id"].(string)
|
||||||
|
require.True(t, ok)
|
||||||
|
|
||||||
|
rr = apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps/"+appID, "")
|
||||||
assert.Equal(t, http.StatusOK, rr.Code)
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
var resp map[string]any
|
var app map[string]any
|
||||||
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &resp))
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &app))
|
||||||
assert.Equal(t, "my-app", resp["name"])
|
assert.Equal(t, "my-app", app["name"])
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestAPIGetAppNotFound(t *testing.T) {
|
func TestAPIGetAppNotFound(t *testing.T) {
|
||||||
@@ -212,7 +249,29 @@ func TestAPIGetAppNotFound(t *testing.T) {
|
|||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/apps/nonexistent")
|
rr := apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps/nonexistent", "")
|
||||||
|
assert.Equal(t, http.StatusNotFound, rr.Code)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestAPIDeleteApp(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
|
body := `{"name":"delete-me","repoUrl":"https://github.com/example/repo"}`
|
||||||
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
|
require.Equal(t, http.StatusCreated, rr.Code)
|
||||||
|
|
||||||
|
var created map[string]any
|
||||||
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &created))
|
||||||
|
|
||||||
|
appID, ok := created["id"].(string)
|
||||||
|
require.True(t, ok)
|
||||||
|
|
||||||
|
rr = apiRequest(t, tc, cookies, http.MethodDelete, "/api/v1/apps/"+appID, "")
|
||||||
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
|
rr = apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps/"+appID, "")
|
||||||
assert.Equal(t, http.StatusNotFound, rr.Code)
|
assert.Equal(t, http.StatusNotFound, rr.Code)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -221,13 +280,17 @@ func TestAPIListDeployments(t *testing.T) {
|
|||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
created, err := tc.appSvc.CreateApp(t.Context(), app.CreateAppInput{
|
body := `{"name":"deploy-app","repoUrl":"https://github.com/example/repo"}`
|
||||||
Name: "deploy-app",
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
RepoURL: "https://github.com/example/repo",
|
require.Equal(t, http.StatusCreated, rr.Code)
|
||||||
})
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/apps/"+created.ID+"/deployments")
|
var created map[string]any
|
||||||
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &created))
|
||||||
|
|
||||||
|
appID, ok := created["id"].(string)
|
||||||
|
require.True(t, ok)
|
||||||
|
|
||||||
|
rr = apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps/"+appID+"/deployments", "")
|
||||||
assert.Equal(t, http.StatusOK, rr.Code)
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
var deployments []any
|
var deployments []any
|
||||||
|
|||||||
@@ -77,14 +77,6 @@ func (h *Handlers) HandleAppCreate() http.HandlerFunc { //nolint:funlen // valid
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
repoURLErr := validateRepoURL(repoURL)
|
|
||||||
if repoURLErr != nil {
|
|
||||||
data["Error"] = "Invalid repository URL: " + repoURLErr.Error()
|
|
||||||
h.renderTemplate(writer, tmpl, "app_new.html", data)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if branch == "" {
|
if branch == "" {
|
||||||
branch = "main"
|
branch = "main"
|
||||||
}
|
}
|
||||||
@@ -233,17 +225,6 @@ func (h *Handlers) HandleAppUpdate() http.HandlerFunc { //nolint:funlen // valid
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
repoURLErr := validateRepoURL(request.FormValue("repo_url"))
|
|
||||||
if repoURLErr != nil {
|
|
||||||
data := h.addGlobals(map[string]any{
|
|
||||||
"App": application,
|
|
||||||
"Error": "Invalid repository URL: " + repoURLErr.Error(),
|
|
||||||
}, request)
|
|
||||||
_ = tmpl.ExecuteTemplate(writer, "app_edit.html", data)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
application.Name = newName
|
application.Name = newName
|
||||||
application.RepoURL = request.FormValue("repo_url")
|
application.RepoURL = request.FormValue("repo_url")
|
||||||
application.Branch = request.FormValue("branch")
|
application.Branch = request.FormValue("branch")
|
||||||
@@ -518,7 +499,7 @@ func (h *Handlers) HandleAppLogs() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
_, _ = writer.Write([]byte(SanitizeLogs(logs))) // #nosec G705 -- logs sanitized, Content-Type is text/plain
|
_, _ = writer.Write([]byte(logs))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -553,7 +534,7 @@ func (h *Handlers) HandleDeploymentLogsAPI() http.HandlerFunc {
|
|||||||
|
|
||||||
logs := ""
|
logs := ""
|
||||||
if deployment.Logs.Valid {
|
if deployment.Logs.Valid {
|
||||||
logs = SanitizeLogs(deployment.Logs.String)
|
logs = deployment.Logs.String
|
||||||
}
|
}
|
||||||
|
|
||||||
response := map[string]any{
|
response := map[string]any{
|
||||||
@@ -600,8 +581,8 @@ func (h *Handlers) HandleDeploymentLogDownload() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if file exists — logPath is constructed internally, not from user input
|
// Check if file exists
|
||||||
_, err := os.Stat(logPath) // #nosec G703 -- path from internal GetLogFilePath, not user input
|
_, err := os.Stat(logPath)
|
||||||
if os.IsNotExist(err) {
|
if os.IsNotExist(err) {
|
||||||
http.NotFound(writer, request)
|
http.NotFound(writer, request)
|
||||||
|
|
||||||
@@ -680,7 +661,7 @@ func (h *Handlers) HandleContainerLogsAPI() http.HandlerFunc {
|
|||||||
}
|
}
|
||||||
|
|
||||||
response := map[string]any{
|
response := map[string]any{
|
||||||
"logs": SanitizeLogs(logs),
|
"logs": logs,
|
||||||
"status": status,
|
"status": status,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -916,7 +897,7 @@ func (h *Handlers) HandleEnvVarAdd() http.HandlerFunc {
|
|||||||
func (h *Handlers) HandleEnvVarDelete() http.HandlerFunc {
|
func (h *Handlers) HandleEnvVarDelete() http.HandlerFunc {
|
||||||
return func(writer http.ResponseWriter, request *http.Request) {
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
appID := chi.URLParam(request, "id")
|
appID := chi.URLParam(request, "id")
|
||||||
envVarIDStr := chi.URLParam(request, "varID")
|
envVarIDStr := chi.URLParam(request, "envID")
|
||||||
|
|
||||||
envVarID, parseErr := strconv.ParseInt(envVarIDStr, 10, 64)
|
envVarID, parseErr := strconv.ParseInt(envVarIDStr, 10, 64)
|
||||||
if parseErr != nil {
|
if parseErr != nil {
|
||||||
@@ -1022,14 +1003,6 @@ func (h *Handlers) HandleVolumeAdd() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
pathErr := validateVolumePaths(hostPath, containerPath)
|
|
||||||
if pathErr != nil {
|
|
||||||
h.log.Error("invalid volume path", "error", pathErr)
|
|
||||||
http.Redirect(writer, request, "/apps/"+application.ID, http.StatusSeeOther)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
volume := models.NewVolume(h.db)
|
volume := models.NewVolume(h.db)
|
||||||
volume.AppID = application.ID
|
volume.AppID = application.ID
|
||||||
volume.HostPath = hostPath
|
volume.HostPath = hostPath
|
||||||
|
|||||||
@@ -1,6 +0,0 @@
|
|||||||
package handlers
|
|
||||||
|
|
||||||
// ValidateRepoURLForTest exports validateRepoURL for testing.
|
|
||||||
func ValidateRepoURLForTest(repoURL string) error {
|
|
||||||
return validateRepoURL(repoURL)
|
|
||||||
}
|
|
||||||
@@ -564,7 +564,7 @@ func TestDeleteEnvVarOwnershipVerification(t *testing.T) { //nolint:dupl // inte
|
|||||||
return "/apps/" + appID + "/env/" + strconv.FormatInt(resourceID, 10) + "/delete"
|
return "/apps/" + appID + "/env/" + strconv.FormatInt(resourceID, 10) + "/delete"
|
||||||
},
|
},
|
||||||
chiParams: func(appID string, resourceID int64) map[string]string {
|
chiParams: func(appID string, resourceID int64) map[string]string {
|
||||||
return map[string]string{"id": appID, "varID": strconv.FormatInt(resourceID, 10)}
|
return map[string]string{"id": appID, "envID": strconv.FormatInt(resourceID, 10)}
|
||||||
},
|
},
|
||||||
handler: func(h *handlers.Handlers) http.HandlerFunc { return h.HandleEnvVarDelete() },
|
handler: func(h *handlers.Handlers) http.HandlerFunc { return h.HandleEnvVarDelete() },
|
||||||
verifyFn: func(t *testing.T, tc *testContext, resourceID int64) {
|
verifyFn: func(t *testing.T, tc *testContext, resourceID int64) {
|
||||||
@@ -695,153 +695,6 @@ func TestDeletePortOwnershipVerification(t *testing.T) {
|
|||||||
assert.NotNil(t, found, "port should still exist after IDOR attempt")
|
assert.NotNil(t, found, "port should still exist after IDOR attempt")
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestHandleEnvVarDeleteUsesCorrectRouteParam verifies that HandleEnvVarDelete
|
|
||||||
// reads the "varID" chi URL parameter (matching the route definition {varID}),
|
|
||||||
// not a mismatched name like "envID".
|
|
||||||
func TestHandleEnvVarDeleteUsesCorrectRouteParam(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
|
|
||||||
createdApp := createTestApp(t, testCtx, "envdelete-param-app")
|
|
||||||
|
|
||||||
envVar := models.NewEnvVar(testCtx.database)
|
|
||||||
envVar.AppID = createdApp.ID
|
|
||||||
envVar.Key = "DELETE_ME"
|
|
||||||
envVar.Value = "gone"
|
|
||||||
require.NoError(t, envVar.Save(context.Background()))
|
|
||||||
|
|
||||||
// Use chi router with the real route pattern to test param name
|
|
||||||
r := chi.NewRouter()
|
|
||||||
r.Post("/apps/{id}/env-vars/{varID}/delete", testCtx.handlers.HandleEnvVarDelete())
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/"+createdApp.ID+"/env-vars/"+strconv.FormatInt(envVar.ID, 10)+"/delete",
|
|
||||||
nil,
|
|
||||||
)
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
|
|
||||||
r.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusSeeOther, recorder.Code)
|
|
||||||
|
|
||||||
// Verify the env var was actually deleted
|
|
||||||
found, findErr := models.FindEnvVar(context.Background(), testCtx.database, envVar.ID)
|
|
||||||
require.NoError(t, findErr)
|
|
||||||
assert.Nil(t, found, "env var should be deleted when using correct route param")
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHandleVolumeAddValidatesPaths verifies that HandleVolumeAdd validates
|
|
||||||
// host and container paths (same as HandleVolumeEdit).
|
|
||||||
func TestHandleVolumeAddValidatesPaths(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
|
|
||||||
createdApp := createTestApp(t, testCtx, "volume-validate-app")
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
hostPath string
|
|
||||||
containerPath string
|
|
||||||
shouldCreate bool
|
|
||||||
}{
|
|
||||||
{"relative host path rejected", "relative/path", "/container", false},
|
|
||||||
{"relative container path rejected", "/host", "relative/path", false},
|
|
||||||
{"unclean host path rejected", "/host/../etc", "/container", false},
|
|
||||||
{"valid paths accepted", "/host/data", "/container/data", true},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
form := url.Values{}
|
|
||||||
form.Set("host_path", tt.hostPath)
|
|
||||||
form.Set("container_path", tt.containerPath)
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/"+createdApp.ID+"/volumes",
|
|
||||||
strings.NewReader(form.Encode()),
|
|
||||||
)
|
|
||||||
request.Header.Set("Content-Type", "application/x-www-form-urlencoded")
|
|
||||||
request = addChiURLParams(request, map[string]string{"id": createdApp.ID})
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
|
|
||||||
handler := testCtx.handlers.HandleVolumeAdd()
|
|
||||||
handler.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusSeeOther, recorder.Code)
|
|
||||||
|
|
||||||
// Check if volume was created by listing volumes
|
|
||||||
volumes, _ := createdApp.GetVolumes(context.Background())
|
|
||||||
found := false
|
|
||||||
|
|
||||||
for _, v := range volumes {
|
|
||||||
if v.HostPath == tt.hostPath && v.ContainerPath == tt.containerPath {
|
|
||||||
found = true
|
|
||||||
// Clean up for isolation
|
|
||||||
_ = v.Delete(context.Background())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if tt.shouldCreate {
|
|
||||||
assert.True(t, found, "volume should be created for valid paths")
|
|
||||||
} else {
|
|
||||||
assert.False(t, found, "volume should NOT be created for invalid paths")
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestSetupRequiredExemptsHealthAndStaticAndAPI verifies that the SetupRequired
|
|
||||||
// middleware allows /health, /s/*, and /api/* paths through even when setup is required.
|
|
||||||
func TestSetupRequiredExemptsHealthAndStaticAndAPI(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
|
|
||||||
// No user created, so setup IS required
|
|
||||||
mw := testCtx.middleware.SetupRequired()
|
|
||||||
|
|
||||||
okHandler := http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
|
|
||||||
w.WriteHeader(http.StatusOK)
|
|
||||||
_, _ = w.Write([]byte("OK"))
|
|
||||||
})
|
|
||||||
|
|
||||||
wrapped := mw(okHandler)
|
|
||||||
|
|
||||||
exemptPaths := []string{"/health", "/s/style.css", "/s/js/app.js", "/api/v1/apps", "/api/v1/login"}
|
|
||||||
|
|
||||||
for _, path := range exemptPaths {
|
|
||||||
t.Run(path, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, path, nil)
|
|
||||||
rr := httptest.NewRecorder()
|
|
||||||
wrapped.ServeHTTP(rr, req)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusOK, rr.Code,
|
|
||||||
"path %s should be exempt from setup redirect", path)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
// Non-exempt path should redirect to /setup
|
|
||||||
t.Run("non-exempt redirects", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, "/", nil)
|
|
||||||
rr := httptest.NewRecorder()
|
|
||||||
wrapped.ServeHTTP(rr, req)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusSeeOther, rr.Code)
|
|
||||||
assert.Equal(t, "/setup", rr.Header().Get("Location"))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestHandleCancelDeployRedirects(t *testing.T) {
|
func TestHandleCancelDeployRedirects(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
|
|||||||
@@ -1,77 +0,0 @@
|
|||||||
package handlers
|
|
||||||
|
|
||||||
import (
|
|
||||||
"errors"
|
|
||||||
"net/url"
|
|
||||||
"regexp"
|
|
||||||
"strings"
|
|
||||||
)
|
|
||||||
|
|
||||||
// Repo URL validation errors.
|
|
||||||
var (
|
|
||||||
errRepoURLEmpty = errors.New("repository URL must not be empty")
|
|
||||||
errRepoURLScheme = errors.New("file:// URLs are not allowed for security reasons")
|
|
||||||
errRepoURLInvalid = errors.New("repository URL must use https://, http://, ssh://, git://, or git@host:path format")
|
|
||||||
errRepoURLNoHost = errors.New("repository URL must include a host")
|
|
||||||
errRepoURLNoPath = errors.New("repository URL must include a path")
|
|
||||||
)
|
|
||||||
|
|
||||||
// scpLikeRepoRe matches SCP-like git URLs: git@host:path (e.g. git@github.com:user/repo.git).
|
|
||||||
// Only the "git" user is allowed, as that is the standard for SSH deploy keys.
|
|
||||||
var scpLikeRepoRe = regexp.MustCompile(`^git@[a-zA-Z0-9._-]+:.+$`)
|
|
||||||
|
|
||||||
// allowedRepoSchemes lists the URL schemes accepted for repository URLs.
|
|
||||||
//
|
|
||||||
//nolint:gochecknoglobals // package-level constant map parsed once
|
|
||||||
var allowedRepoSchemes = map[string]bool{
|
|
||||||
"https": true,
|
|
||||||
"http": true,
|
|
||||||
"ssh": true,
|
|
||||||
"git": true,
|
|
||||||
}
|
|
||||||
|
|
||||||
// validateRepoURL checks that the given repository URL is valid and uses an allowed scheme.
|
|
||||||
func validateRepoURL(repoURL string) error {
|
|
||||||
if strings.TrimSpace(repoURL) == "" {
|
|
||||||
return errRepoURLEmpty
|
|
||||||
}
|
|
||||||
|
|
||||||
// Reject path traversal in any URL format
|
|
||||||
if strings.Contains(repoURL, "..") {
|
|
||||||
return errRepoURLInvalid
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check for SCP-like git URLs first (git@host:path)
|
|
||||||
if scpLikeRepoRe.MatchString(repoURL) {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Reject file:// explicitly
|
|
||||||
if strings.HasPrefix(strings.ToLower(repoURL), "file://") {
|
|
||||||
return errRepoURLScheme
|
|
||||||
}
|
|
||||||
|
|
||||||
return validateParsedRepoURL(repoURL)
|
|
||||||
}
|
|
||||||
|
|
||||||
// validateParsedRepoURL validates a standard URL-format repository URL.
|
|
||||||
func validateParsedRepoURL(repoURL string) error {
|
|
||||||
parsed, err := url.Parse(repoURL)
|
|
||||||
if err != nil {
|
|
||||||
return errRepoURLInvalid
|
|
||||||
}
|
|
||||||
|
|
||||||
if !allowedRepoSchemes[strings.ToLower(parsed.Scheme)] {
|
|
||||||
return errRepoURLInvalid
|
|
||||||
}
|
|
||||||
|
|
||||||
if parsed.Host == "" {
|
|
||||||
return errRepoURLNoHost
|
|
||||||
}
|
|
||||||
|
|
||||||
if parsed.Path == "" || parsed.Path == "/" {
|
|
||||||
return errRepoURLNoPath
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
@@ -1,60 +0,0 @@
|
|||||||
package handlers_test
|
|
||||||
|
|
||||||
import (
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/handlers"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestValidateRepoURL(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
url string
|
|
||||||
wantErr bool
|
|
||||||
}{
|
|
||||||
// Valid URLs
|
|
||||||
{name: "https URL", url: "https://github.com/user/repo.git", wantErr: false},
|
|
||||||
{name: "http URL", url: "http://github.com/user/repo.git", wantErr: false},
|
|
||||||
{name: "ssh URL", url: "ssh://git@github.com/user/repo.git", wantErr: false},
|
|
||||||
{name: "git URL", url: "git://github.com/user/repo.git", wantErr: false},
|
|
||||||
{name: "SCP-like URL", url: "git@github.com:user/repo.git", wantErr: false},
|
|
||||||
{name: "SCP-like with dots", url: "git@git.example.com:org/repo.git", wantErr: false},
|
|
||||||
{name: "https without .git", url: "https://github.com/user/repo", wantErr: false},
|
|
||||||
{name: "https with port", url: "https://git.example.com:8443/user/repo.git", wantErr: false},
|
|
||||||
|
|
||||||
// Invalid URLs
|
|
||||||
{name: "empty string", url: "", wantErr: true},
|
|
||||||
{name: "whitespace only", url: " ", wantErr: true},
|
|
||||||
{name: "file URL", url: "file:///etc/passwd", wantErr: true},
|
|
||||||
{name: "file URL uppercase", url: "FILE:///etc/passwd", wantErr: true},
|
|
||||||
{name: "bare path", url: "/some/local/path", wantErr: true},
|
|
||||||
{name: "relative path", url: "../repo", wantErr: true},
|
|
||||||
{name: "just a word", url: "notaurl", wantErr: true},
|
|
||||||
{name: "ftp URL", url: "ftp://example.com/repo.git", wantErr: true},
|
|
||||||
{name: "no host https", url: "https:///path", wantErr: true},
|
|
||||||
{name: "no path https", url: "https://github.com", wantErr: true},
|
|
||||||
{name: "no path https trailing slash", url: "https://github.com/", wantErr: true},
|
|
||||||
{name: "SCP-like non-git user", url: "root@github.com:user/repo.git", wantErr: true},
|
|
||||||
{name: "SCP-like arbitrary user", url: "admin@github.com:user/repo.git", wantErr: true},
|
|
||||||
{name: "path traversal SCP", url: "git@github.com:../../etc/passwd", wantErr: true},
|
|
||||||
{name: "path traversal https", url: "https://github.com/user/../../../etc/passwd", wantErr: true},
|
|
||||||
{name: "path traversal in middle", url: "https://github.com/user/repo/../secret", wantErr: true},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tc := range tests {
|
|
||||||
t.Run(tc.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
err := handlers.ValidateRepoURLForTest(tc.url)
|
|
||||||
if tc.wantErr && err == nil {
|
|
||||||
t.Errorf("ValidateRepoURLForTest(%q) = nil, want error", tc.url)
|
|
||||||
}
|
|
||||||
|
|
||||||
if !tc.wantErr && err != nil {
|
|
||||||
t.Errorf("ValidateRepoURLForTest(%q) = %v, want nil", tc.url, err)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,30 +0,0 @@
|
|||||||
package handlers
|
|
||||||
|
|
||||||
import (
|
|
||||||
"regexp"
|
|
||||||
"strings"
|
|
||||||
)
|
|
||||||
|
|
||||||
// ansiEscapePattern matches ANSI escape sequences (CSI, OSC, and single-character escapes).
|
|
||||||
var ansiEscapePattern = regexp.MustCompile(`(\x1b\[[0-9;]*[a-zA-Z]|\x1b\][^\x07]*\x07|\x1b[^[\]])`)
|
|
||||||
|
|
||||||
// SanitizeLogs strips ANSI escape sequences and non-printable control characters
|
|
||||||
// from container log output. Newlines (\n), carriage returns (\r), and tabs (\t)
|
|
||||||
// are preserved. This ensures that attacker-controlled container output cannot
|
|
||||||
// inject terminal escape sequences or other dangerous control characters.
|
|
||||||
func SanitizeLogs(input string) string {
|
|
||||||
// Strip ANSI escape sequences
|
|
||||||
result := ansiEscapePattern.ReplaceAllString(input, "")
|
|
||||||
|
|
||||||
// Strip remaining non-printable characters (keep \n, \r, \t)
|
|
||||||
var b strings.Builder
|
|
||||||
b.Grow(len(result))
|
|
||||||
|
|
||||||
for _, r := range result {
|
|
||||||
if r == '\n' || r == '\r' || r == '\t' || r >= ' ' {
|
|
||||||
b.WriteRune(r)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return b.String()
|
|
||||||
}
|
|
||||||
@@ -1,84 +0,0 @@
|
|||||||
package handlers_test
|
|
||||||
|
|
||||||
import (
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/handlers"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestSanitizeLogs(t *testing.T) { //nolint:funlen // table-driven tests
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
input string
|
|
||||||
expected string
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "plain text unchanged",
|
|
||||||
input: "hello world\n",
|
|
||||||
expected: "hello world\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips ANSI color codes",
|
|
||||||
input: "\x1b[31mERROR\x1b[0m: something failed\n",
|
|
||||||
expected: "ERROR: something failed\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips OSC sequences",
|
|
||||||
input: "\x1b]0;window title\x07normal text\n",
|
|
||||||
expected: "normal text\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips null bytes",
|
|
||||||
input: "hello\x00world\n",
|
|
||||||
expected: "helloworld\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips bell characters",
|
|
||||||
input: "alert\x07here\n",
|
|
||||||
expected: "alerthere\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "preserves tabs",
|
|
||||||
input: "field1\tfield2\tfield3\n",
|
|
||||||
expected: "field1\tfield2\tfield3\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "preserves carriage returns",
|
|
||||||
input: "line1\r\nline2\r\n",
|
|
||||||
expected: "line1\r\nline2\r\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips mixed escape sequences",
|
|
||||||
input: "\x1b[32m2024-01-01\x1b[0m \x1b[1mINFO\x1b[0m starting\x00\n",
|
|
||||||
expected: "2024-01-01 INFO starting\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "empty string",
|
|
||||||
input: "",
|
|
||||||
expected: "",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "only control characters",
|
|
||||||
input: "\x00\x01\x02\x03",
|
|
||||||
expected: "",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "cursor movement sequences stripped",
|
|
||||||
input: "\x1b[2J\x1b[H\x1b[3Atext\n",
|
|
||||||
expected: "text\n",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
got := handlers.SanitizeLogs(tt.input)
|
|
||||||
if got != tt.expected {
|
|
||||||
t.Errorf("SanitizeLogs(%q) = %q, want %q", tt.input, got, tt.expected)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,81 +0,0 @@
|
|||||||
package middleware //nolint:testpackage // tests internal CORS behavior
|
|
||||||
|
|
||||||
import (
|
|
||||||
"log/slog"
|
|
||||||
"net/http"
|
|
||||||
"net/http/httptest"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/config"
|
|
||||||
)
|
|
||||||
|
|
||||||
//nolint:gosec // test credentials
|
|
||||||
func newCORSTestMiddleware(corsOrigins string) *Middleware {
|
|
||||||
return &Middleware{
|
|
||||||
log: slog.Default(),
|
|
||||||
params: &Params{
|
|
||||||
Config: &config.Config{
|
|
||||||
CORSOrigins: corsOrigins,
|
|
||||||
SessionSecret: "test-secret-32-bytes-long-enough",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestCORS_NoOriginsConfigured_NoCORSHeaders(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
m := newCORSTestMiddleware("")
|
|
||||||
handler := m.CORS()(http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
|
|
||||||
w.WriteHeader(http.StatusOK)
|
|
||||||
}))
|
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, "/", nil)
|
|
||||||
req.Header.Set("Origin", "https://evil.com")
|
|
||||||
|
|
||||||
rec := httptest.NewRecorder()
|
|
||||||
handler.ServeHTTP(rec, req)
|
|
||||||
|
|
||||||
assert.Empty(t, rec.Header().Get("Access-Control-Allow-Origin"),
|
|
||||||
"expected no CORS headers when no origins configured")
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestCORS_OriginsConfigured_AllowsMatchingOrigin(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
m := newCORSTestMiddleware("https://app.example.com,https://other.example.com")
|
|
||||||
handler := m.CORS()(http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
|
|
||||||
w.WriteHeader(http.StatusOK)
|
|
||||||
}))
|
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, "/", nil)
|
|
||||||
req.Header.Set("Origin", "https://app.example.com")
|
|
||||||
|
|
||||||
rec := httptest.NewRecorder()
|
|
||||||
handler.ServeHTTP(rec, req)
|
|
||||||
|
|
||||||
assert.Equal(t, "https://app.example.com",
|
|
||||||
rec.Header().Get("Access-Control-Allow-Origin"))
|
|
||||||
assert.Equal(t, "true",
|
|
||||||
rec.Header().Get("Access-Control-Allow-Credentials"))
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestCORS_OriginsConfigured_RejectsNonMatchingOrigin(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
m := newCORSTestMiddleware("https://app.example.com")
|
|
||||||
handler := m.CORS()(http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
|
|
||||||
w.WriteHeader(http.StatusOK)
|
|
||||||
}))
|
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, "/", nil)
|
|
||||||
req.Header.Set("Origin", "https://evil.com")
|
|
||||||
|
|
||||||
rec := httptest.NewRecorder()
|
|
||||||
handler.ServeHTTP(rec, req)
|
|
||||||
|
|
||||||
assert.Empty(t, rec.Header().Get("Access-Control-Allow-Origin"),
|
|
||||||
"expected no CORS headers for non-matching origin")
|
|
||||||
}
|
|
||||||
@@ -177,48 +177,17 @@ func realIP(r *http.Request) string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// CORS returns CORS middleware.
|
// CORS returns CORS middleware.
|
||||||
// When UPAAS_CORS_ORIGINS is empty (default), no CORS headers are sent
|
|
||||||
// (same-origin only). When configured, only the specified origins are
|
|
||||||
// allowed and credentials (cookies) are permitted.
|
|
||||||
func (m *Middleware) CORS() func(http.Handler) http.Handler {
|
func (m *Middleware) CORS() func(http.Handler) http.Handler {
|
||||||
origins := parseCORSOrigins(m.params.Config.CORSOrigins)
|
|
||||||
|
|
||||||
// No origins configured — no CORS headers (same-origin policy).
|
|
||||||
if len(origins) == 0 {
|
|
||||||
return func(next http.Handler) http.Handler {
|
|
||||||
return next
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return cors.Handler(cors.Options{
|
return cors.Handler(cors.Options{
|
||||||
AllowedOrigins: origins,
|
AllowedOrigins: []string{"*"},
|
||||||
AllowedMethods: []string{"GET", "POST", "PUT", "DELETE", "OPTIONS"},
|
AllowedMethods: []string{"GET", "POST", "PUT", "DELETE", "OPTIONS"},
|
||||||
AllowedHeaders: []string{"Accept", "Authorization", "Content-Type", "X-CSRF-Token"},
|
AllowedHeaders: []string{"Accept", "Authorization", "Content-Type", "X-CSRF-Token"},
|
||||||
ExposedHeaders: []string{"Link"},
|
ExposedHeaders: []string{"Link"},
|
||||||
AllowCredentials: true,
|
AllowCredentials: false,
|
||||||
MaxAge: corsMaxAge,
|
MaxAge: corsMaxAge,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
// parseCORSOrigins splits a comma-separated origin string into a slice,
|
|
||||||
// trimming whitespace. Returns nil if the input is empty.
|
|
||||||
func parseCORSOrigins(raw string) []string {
|
|
||||||
if raw == "" {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
parts := strings.Split(raw, ",")
|
|
||||||
origins := make([]string, 0, len(parts))
|
|
||||||
|
|
||||||
for _, p := range parts {
|
|
||||||
if o := strings.TrimSpace(p); o != "" {
|
|
||||||
origins = append(origins, o)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return origins
|
|
||||||
}
|
|
||||||
|
|
||||||
// MetricsAuth returns basic auth middleware for metrics endpoint.
|
// MetricsAuth returns basic auth middleware for metrics endpoint.
|
||||||
func (m *Middleware) MetricsAuth() func(http.Handler) http.Handler {
|
func (m *Middleware) MetricsAuth() func(http.Handler) http.Handler {
|
||||||
if m.params.Config.MetricsUsername == "" {
|
if m.params.Config.MetricsUsername == "" {
|
||||||
@@ -411,14 +380,8 @@ func (m *Middleware) SetupRequired() func(http.Handler) http.Handler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if setupRequired {
|
if setupRequired {
|
||||||
path := request.URL.Path
|
// Allow access to setup page
|
||||||
|
if request.URL.Path == "/setup" {
|
||||||
// Allow access to setup page, health endpoint, static
|
|
||||||
// assets, and API routes even before setup is complete.
|
|
||||||
if path == "/setup" ||
|
|
||||||
path == "/health" ||
|
|
||||||
strings.HasPrefix(path, "/s/") ||
|
|
||||||
strings.HasPrefix(path, "/api/") {
|
|
||||||
next.ServeHTTP(writer, request)
|
next.ServeHTTP(writer, request)
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|||||||
@@ -114,7 +114,10 @@ func (s *Server) SetupRoutes() {
|
|||||||
r.Get("/whoami", s.handlers.HandleAPIWhoAmI())
|
r.Get("/whoami", s.handlers.HandleAPIWhoAmI())
|
||||||
|
|
||||||
r.Get("/apps", s.handlers.HandleAPIListApps())
|
r.Get("/apps", s.handlers.HandleAPIListApps())
|
||||||
|
r.Post("/apps", s.handlers.HandleAPICreateApp())
|
||||||
r.Get("/apps/{id}", s.handlers.HandleAPIGetApp())
|
r.Get("/apps/{id}", s.handlers.HandleAPIGetApp())
|
||||||
|
r.Delete("/apps/{id}", s.handlers.HandleAPIDeleteApp())
|
||||||
|
r.Post("/apps/{id}/deploy", s.handlers.HandleAPITriggerDeploy())
|
||||||
r.Get("/apps/{id}/deployments", s.handlers.HandleAPIListDeployments())
|
r.Get("/apps/{id}/deployments", s.handlers.HandleAPIListDeployments())
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -11,7 +11,6 @@ import (
|
|||||||
"log/slog"
|
"log/slog"
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
@@ -473,7 +472,7 @@ func (svc *Service) runBuildAndDeploy(
|
|||||||
// Build phase with timeout
|
// Build phase with timeout
|
||||||
imageID, err := svc.buildImageWithTimeout(deployCtx, app, deployment)
|
imageID, err := svc.buildImageWithTimeout(deployCtx, app, deployment)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment, "")
|
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment)
|
||||||
if cancelErr != nil {
|
if cancelErr != nil {
|
||||||
return cancelErr
|
return cancelErr
|
||||||
}
|
}
|
||||||
@@ -486,7 +485,7 @@ func (svc *Service) runBuildAndDeploy(
|
|||||||
// Deploy phase with timeout
|
// Deploy phase with timeout
|
||||||
err = svc.deployContainerWithTimeout(deployCtx, app, deployment, imageID)
|
err = svc.deployContainerWithTimeout(deployCtx, app, deployment, imageID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment, imageID)
|
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment)
|
||||||
if cancelErr != nil {
|
if cancelErr != nil {
|
||||||
return cancelErr
|
return cancelErr
|
||||||
}
|
}
|
||||||
@@ -662,77 +661,24 @@ func (svc *Service) cancelActiveDeploy(appID string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// checkCancelled checks if the deploy context was cancelled (by a newer deploy)
|
// checkCancelled checks if the deploy context was cancelled (by a newer deploy)
|
||||||
// and if so, marks the deployment as cancelled and cleans up orphan resources.
|
// and if so, marks the deployment as cancelled. Returns ErrDeployCancelled or nil.
|
||||||
// Returns ErrDeployCancelled or nil.
|
|
||||||
func (svc *Service) checkCancelled(
|
func (svc *Service) checkCancelled(
|
||||||
deployCtx context.Context,
|
deployCtx context.Context,
|
||||||
bgCtx context.Context,
|
bgCtx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
imageID string,
|
|
||||||
) error {
|
) error {
|
||||||
if !errors.Is(deployCtx.Err(), context.Canceled) {
|
if !errors.Is(deployCtx.Err(), context.Canceled) {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
svc.log.Info("deployment cancelled", "app", app.Name)
|
svc.log.Info("deployment cancelled by newer deploy", "app", app.Name)
|
||||||
|
|
||||||
svc.cleanupCancelledDeploy(bgCtx, app, deployment, imageID)
|
|
||||||
|
|
||||||
_ = deployment.MarkFinished(bgCtx, models.DeploymentStatusCancelled)
|
_ = deployment.MarkFinished(bgCtx, models.DeploymentStatusCancelled)
|
||||||
|
|
||||||
return ErrDeployCancelled
|
return ErrDeployCancelled
|
||||||
}
|
}
|
||||||
|
|
||||||
// cleanupCancelledDeploy removes orphan resources left by a cancelled deployment.
|
|
||||||
func (svc *Service) cleanupCancelledDeploy(
|
|
||||||
ctx context.Context,
|
|
||||||
app *models.App,
|
|
||||||
deployment *models.Deployment,
|
|
||||||
imageID string,
|
|
||||||
) {
|
|
||||||
// Clean up the intermediate Docker image if one was built
|
|
||||||
if imageID != "" {
|
|
||||||
removeErr := svc.docker.RemoveImage(ctx, imageID)
|
|
||||||
if removeErr != nil {
|
|
||||||
svc.log.Error("failed to remove image from cancelled deploy",
|
|
||||||
"error", removeErr, "app", app.Name, "image", imageID)
|
|
||||||
_ = deployment.AppendLog(ctx, "WARNING: failed to clean up image "+imageID+": "+removeErr.Error())
|
|
||||||
} else {
|
|
||||||
svc.log.Info("cleaned up image from cancelled deploy",
|
|
||||||
"app", app.Name, "image", imageID)
|
|
||||||
_ = deployment.AppendLog(ctx, "Cleaned up intermediate image: "+imageID)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Clean up the build directory for this deployment
|
|
||||||
buildDir := svc.GetBuildDir(app.Name)
|
|
||||||
|
|
||||||
entries, err := os.ReadDir(buildDir)
|
|
||||||
if err != nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
prefix := fmt.Sprintf("%d-", deployment.ID)
|
|
||||||
|
|
||||||
for _, entry := range entries {
|
|
||||||
if entry.IsDir() && strings.HasPrefix(entry.Name(), prefix) {
|
|
||||||
dirPath := filepath.Join(buildDir, entry.Name())
|
|
||||||
|
|
||||||
removeErr := os.RemoveAll(dirPath)
|
|
||||||
if removeErr != nil {
|
|
||||||
svc.log.Error("failed to remove build dir from cancelled deploy",
|
|
||||||
"error", removeErr, "path", dirPath)
|
|
||||||
} else {
|
|
||||||
svc.log.Info("cleaned up build dir from cancelled deploy",
|
|
||||||
"app", app.Name, "path", dirPath)
|
|
||||||
|
|
||||||
_ = deployment.AppendLog(ctx, "Cleaned up build directory")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (svc *Service) fetchWebhookEvent(
|
func (svc *Service) fetchWebhookEvent(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
webhookEventID *int64,
|
webhookEventID *int64,
|
||||||
|
|||||||
@@ -1,63 +0,0 @@
|
|||||||
package deploy_test
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"log/slog"
|
|
||||||
"os"
|
|
||||||
"path/filepath"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
|
||||||
"github.com/stretchr/testify/require"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/config"
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestCleanupCancelledDeploy_RemovesBuildDir(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
tmpDir := t.TempDir()
|
|
||||||
cfg := &config.Config{DataDir: tmpDir}
|
|
||||||
|
|
||||||
svc := deploy.NewTestServiceWithConfig(slog.Default(), cfg, nil)
|
|
||||||
|
|
||||||
// Create a fake build directory matching the deployment pattern
|
|
||||||
appName := "test-app"
|
|
||||||
buildDir := svc.GetBuildDirExported(appName)
|
|
||||||
require.NoError(t, os.MkdirAll(buildDir, 0o750))
|
|
||||||
|
|
||||||
// Create deployment-specific dir: <deploymentID>-<random>
|
|
||||||
deployDir := filepath.Join(buildDir, "42-abc123")
|
|
||||||
require.NoError(t, os.MkdirAll(deployDir, 0o750))
|
|
||||||
|
|
||||||
// Create a file inside to verify full removal
|
|
||||||
require.NoError(t, os.WriteFile(filepath.Join(deployDir, "work"), []byte("test"), 0o600))
|
|
||||||
|
|
||||||
// Also create a dir for a different deployment (should NOT be removed)
|
|
||||||
otherDir := filepath.Join(buildDir, "99-xyz789")
|
|
||||||
require.NoError(t, os.MkdirAll(otherDir, 0o750))
|
|
||||||
|
|
||||||
// Run cleanup for deployment 42
|
|
||||||
svc.CleanupCancelledDeploy(context.Background(), appName, 42, "")
|
|
||||||
|
|
||||||
// Deployment 42's dir should be gone
|
|
||||||
_, err := os.Stat(deployDir)
|
|
||||||
assert.True(t, os.IsNotExist(err), "deployment build dir should be removed")
|
|
||||||
|
|
||||||
// Deployment 99's dir should still exist
|
|
||||||
_, err = os.Stat(otherDir)
|
|
||||||
assert.NoError(t, err, "other deployment build dir should not be removed")
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestCleanupCancelledDeploy_NoBuildDir(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
tmpDir := t.TempDir()
|
|
||||||
cfg := &config.Config{DataDir: tmpDir}
|
|
||||||
|
|
||||||
svc := deploy.NewTestServiceWithConfig(slog.Default(), cfg, nil)
|
|
||||||
|
|
||||||
// Should not panic when build dir doesn't exist
|
|
||||||
svc.CleanupCancelledDeploy(context.Background(), "nonexistent-app", 1, "")
|
|
||||||
}
|
|
||||||
@@ -2,14 +2,7 @@ package deploy
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"os"
|
|
||||||
"path/filepath"
|
|
||||||
"strings"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/config"
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/docker"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// NewTestService creates a Service with minimal dependencies for testing.
|
// NewTestService creates a Service with minimal dependencies for testing.
|
||||||
@@ -38,45 +31,3 @@ func (svc *Service) TryLockApp(appID string) bool {
|
|||||||
func (svc *Service) UnlockApp(appID string) {
|
func (svc *Service) UnlockApp(appID string) {
|
||||||
svc.unlockApp(appID)
|
svc.unlockApp(appID)
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewTestServiceWithConfig creates a Service with config and docker client for testing.
|
|
||||||
func NewTestServiceWithConfig(log *slog.Logger, cfg *config.Config, dockerClient *docker.Client) *Service {
|
|
||||||
return &Service{
|
|
||||||
log: log,
|
|
||||||
config: cfg,
|
|
||||||
docker: dockerClient,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// CleanupCancelledDeploy exposes the build directory cleanup portion of
|
|
||||||
// cleanupCancelledDeploy for testing. It removes build directories matching
|
|
||||||
// the deployment ID prefix.
|
|
||||||
func (svc *Service) CleanupCancelledDeploy(
|
|
||||||
_ context.Context,
|
|
||||||
appName string,
|
|
||||||
deploymentID int64,
|
|
||||||
_ string,
|
|
||||||
) {
|
|
||||||
// We can't create real models.App/Deployment in tests easily,
|
|
||||||
// so we test the build dir cleanup portion directly.
|
|
||||||
buildDir := svc.GetBuildDir(appName)
|
|
||||||
|
|
||||||
entries, err := os.ReadDir(buildDir)
|
|
||||||
if err != nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
prefix := fmt.Sprintf("%d-", deploymentID)
|
|
||||||
|
|
||||||
for _, entry := range entries {
|
|
||||||
if entry.IsDir() && strings.HasPrefix(entry.Name(), prefix) {
|
|
||||||
dirPath := filepath.Join(buildDir, entry.Name())
|
|
||||||
_ = os.RemoveAll(dirPath)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// GetBuildDirExported exposes GetBuildDir for testing.
|
|
||||||
func (svc *Service) GetBuildDirExported(appName string) string {
|
|
||||||
return svc.GetBuildDir(appName)
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -10,7 +10,6 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
@@ -248,15 +247,10 @@ func (svc *Service) sendNtfy(
|
|||||||
) error {
|
) error {
|
||||||
svc.log.Debug("sending ntfy notification", "topic", topic, "title", title)
|
svc.log.Debug("sending ntfy notification", "topic", topic, "title", title)
|
||||||
|
|
||||||
parsedURL, err := url.ParseRequestURI(topic)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("invalid ntfy topic URL: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
request, err := http.NewRequestWithContext(
|
request, err := http.NewRequestWithContext(
|
||||||
ctx,
|
ctx,
|
||||||
http.MethodPost,
|
http.MethodPost,
|
||||||
parsedURL.String(),
|
topic,
|
||||||
bytes.NewBufferString(message),
|
bytes.NewBufferString(message),
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -266,7 +260,7 @@ func (svc *Service) sendNtfy(
|
|||||||
request.Header.Set("Title", title)
|
request.Header.Set("Title", title)
|
||||||
request.Header.Set("Priority", svc.ntfyPriority(priority))
|
request.Header.Set("Priority", svc.ntfyPriority(priority))
|
||||||
|
|
||||||
resp, err := svc.client.Do(request) // #nosec G704 -- URL from validated config, not user input
|
resp, err := svc.client.Do(request)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to send ntfy request: %w", err)
|
return fmt.Errorf("failed to send ntfy request: %w", err)
|
||||||
}
|
}
|
||||||
@@ -346,15 +340,10 @@ func (svc *Service) sendSlack(
|
|||||||
return fmt.Errorf("failed to marshal slack payload: %w", err)
|
return fmt.Errorf("failed to marshal slack payload: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
parsedWebhookURL, err := url.ParseRequestURI(webhookURL)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("invalid slack webhook URL: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
request, err := http.NewRequestWithContext(
|
request, err := http.NewRequestWithContext(
|
||||||
ctx,
|
ctx,
|
||||||
http.MethodPost,
|
http.MethodPost,
|
||||||
parsedWebhookURL.String(),
|
webhookURL,
|
||||||
bytes.NewBuffer(body),
|
bytes.NewBuffer(body),
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -363,7 +352,7 @@ func (svc *Service) sendSlack(
|
|||||||
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
request.Header.Set("Content-Type", "application/json")
|
||||||
|
|
||||||
resp, err := svc.client.Do(request) // #nosec G704 -- URL from validated config, not user input
|
resp, err := svc.client.Do(request)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to send slack request: %w", err)
|
return fmt.Errorf("failed to send slack request: %w", err)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ import (
|
|||||||
|
|
||||||
// KeyPair contains an SSH key pair.
|
// KeyPair contains an SSH key pair.
|
||||||
type KeyPair struct {
|
type KeyPair struct {
|
||||||
PrivateKey string `json:"-"`
|
PrivateKey string
|
||||||
PublicKey string
|
PublicKey string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -369,7 +369,7 @@ document.addEventListener("alpine:init", () => {
|
|||||||
init() {
|
init() {
|
||||||
// Read initial logs from script tag (avoids escaping issues)
|
// Read initial logs from script tag (avoids escaping issues)
|
||||||
const initialLogsEl = this.$el.querySelector(".initial-logs");
|
const initialLogsEl = this.$el.querySelector(".initial-logs");
|
||||||
this.logs = initialLogsEl?.dataset.logs || "Loading...";
|
this.logs = initialLogsEl?.textContent || "Loading...";
|
||||||
|
|
||||||
// Set up scroll tracking
|
// Set up scroll tracking
|
||||||
this.$nextTick(() => {
|
this.$nextTick(() => {
|
||||||
|
|||||||
@@ -98,7 +98,7 @@
|
|||||||
title="Scroll to bottom"
|
title="Scroll to bottom"
|
||||||
>↓ Follow</button>
|
>↓ Follow</button>
|
||||||
</div>
|
</div>
|
||||||
{{if .Logs.Valid}}<div hidden class="initial-logs" data-logs="{{.Logs.String}}"></div>{{end}}
|
{{if .Logs.Valid}}<script type="text/plain" class="initial-logs">{{.Logs.String}}</script>{{end}}
|
||||||
</div>
|
</div>
|
||||||
{{end}}
|
{{end}}
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
Reference in New Issue
Block a user