Compare commits
16 Commits
schema-con
...
chore/code
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
4f81d9cb70 | ||
|
|
387a0f1d9a | ||
|
|
1417a87dff | ||
|
|
e2e270a557 | ||
| 8ad2c6e42c | |||
|
|
0fcf12d2cc | ||
| 3a4e999382 | |||
|
|
728b29ef16 | ||
| f61d4d0f91 | |||
|
|
8ec04fdadb | ||
| 06e8e66443 | |||
|
|
95a690e805 | ||
|
|
802518b917 | ||
| b47f871412 | |||
|
|
02847eea92 | ||
|
|
506c795f16 |
180
BUGS.md
180
BUGS.md
@@ -1,180 +0,0 @@
|
|||||||
# Bugs in µPaaS
|
|
||||||
|
|
||||||
## 1. Potential Race Condition in Log Writing
|
|
||||||
|
|
||||||
### Description
|
|
||||||
In the deployment service, when a deployment fails, the `failDeployment` function calls `writeLogsToFile` which may be called concurrently with the async log writer's flush operations. This could lead to partial or corrupted log files.
|
|
||||||
|
|
||||||
### Location
|
|
||||||
`internal/service/deploy/deploy.go:1169` in `failDeployment` function
|
|
||||||
|
|
||||||
### Proposed Fix
|
|
||||||
1. Add synchronization to ensure only one log write operation occurs at a time
|
|
||||||
2. Modify the `deploymentLogWriter` to track completion status and prevent concurrent writes
|
|
||||||
3. Add a wait mechanism in `failDeployment` to ensure any ongoing flush operations complete before writing logs to file
|
|
||||||
|
|
||||||
```go
|
|
||||||
// Add a mutex to deploymentLogWriter
|
|
||||||
type deploymentLogWriter struct {
|
|
||||||
// existing fields...
|
|
||||||
mu sync.Mutex
|
|
||||||
writeMu sync.Mutex // Add this for file writing synchronization
|
|
||||||
done chan struct{}
|
|
||||||
flushed sync.WaitGroup
|
|
||||||
}
|
|
||||||
|
|
||||||
// In writeLogsToFile, ensure exclusive access
|
|
||||||
func (svc *Service) writeLogsToFile(app *models.App, deployment *models.Deployment) {
|
|
||||||
svc.writeMu.Lock() // Add this mutex to Service struct
|
|
||||||
defer svc.writeMu.Unlock()
|
|
||||||
// existing code...
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
## 2. Incomplete Error Handling in Container Operations
|
|
||||||
|
|
||||||
### Description
|
|
||||||
In the Docker client's `performClone` function, if `createGitContainer` fails, the SSH key file created earlier is not cleaned up, causing a potential security risk.
|
|
||||||
|
|
||||||
### Location
|
|
||||||
`internal/docker/client.go:597` in `performClone` function
|
|
||||||
|
|
||||||
### Proposed Fix
|
|
||||||
Add proper cleanup using `defer` immediately after creating the SSH key file:
|
|
||||||
|
|
||||||
```go
|
|
||||||
// After writing SSH key file (line 578)
|
|
||||||
keyFileCreated := false
|
|
||||||
err = os.WriteFile(cfg.keyFile, []byte(cfg.sshPrivateKey), sshKeyPermissions)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("failed to write SSH key: %w", err)
|
|
||||||
}
|
|
||||||
keyFileCreated = true
|
|
||||||
|
|
||||||
defer func() {
|
|
||||||
if keyFileCreated {
|
|
||||||
removeErr := os.Remove(cfg.keyFile)
|
|
||||||
if removeErr != nil {
|
|
||||||
c.log.Error("failed to remove SSH key file", "error", removeErr)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
```
|
|
||||||
|
|
||||||
## 3. Missing Context Cancellation Check During Build
|
|
||||||
|
|
||||||
### Description
|
|
||||||
In the deployment service's `streamBuildOutput` function, long-running Docker build operations may not properly respond to context cancellation, causing deployments to hang even when cancelled.
|
|
||||||
|
|
||||||
### Location
|
|
||||||
`internal/docker/client.go:542` in `streamBuildOutput` function
|
|
||||||
|
|
||||||
### Proposed Fix
|
|
||||||
Add context checking in the scanner loop:
|
|
||||||
|
|
||||||
```go
|
|
||||||
for scanner.Scan() {
|
|
||||||
select {
|
|
||||||
case <-ctx.Done():
|
|
||||||
return ctx.Err()
|
|
||||||
default:
|
|
||||||
}
|
|
||||||
|
|
||||||
line := scanner.Bytes()
|
|
||||||
// existing code...
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
## 4. Inconsistent Container Removal in Error Cases
|
|
||||||
|
|
||||||
### Description
|
|
||||||
When deployment fails during container creation, the already-created container is not removed, leading to orphaned containers that consume resources.
|
|
||||||
|
|
||||||
### Location
|
|
||||||
`internal/service/deploy/deploy.go:969` in `createAndStartContainer` function
|
|
||||||
|
|
||||||
### Proposed Fix
|
|
||||||
Add cleanup of created container on start failure:
|
|
||||||
|
|
||||||
```go
|
|
||||||
containerID, err := svc.docker.CreateContainer(ctx, containerOpts)
|
|
||||||
if err != nil {
|
|
||||||
svc.notify.NotifyDeployFailed(ctx, app, deployment, err)
|
|
||||||
svc.failDeployment(ctx, app, deployment, fmt.Errorf("failed to create container: %w", err))
|
|
||||||
return "", fmt.Errorf("failed to create container: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Add cleanup defer for error cases
|
|
||||||
defer func() {
|
|
||||||
if err != nil {
|
|
||||||
// If we have a container ID but returning an error, clean it up
|
|
||||||
_ = svc.docker.RemoveContainer(context.Background(), containerID, true)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
startErr := svc.docker.StartContainer(ctx, containerID)
|
|
||||||
if startErr != nil {
|
|
||||||
svc.notify.NotifyDeployFailed(ctx, app, deployment, startErr)
|
|
||||||
svc.failDeployment(ctx, app, deployment, fmt.Errorf("failed to start container: %w", startErr))
|
|
||||||
err = startErr // Set err so defer cleanup runs
|
|
||||||
return "", fmt.Errorf("failed to start container: %w", startErr)
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
## 5. Potential Data Race in Active Deployments Tracking
|
|
||||||
|
|
||||||
### Description
|
|
||||||
The `activeDeploys` sync.Map in the deployment service may have race conditions when multiple concurrent deployments try to access the same app's deployment state.
|
|
||||||
|
|
||||||
### Location
|
|
||||||
`internal/service/deploy/deploy.go:226` and related functions
|
|
||||||
|
|
||||||
### Proposed Fix
|
|
||||||
Add proper locking around active deploy operations:
|
|
||||||
|
|
||||||
```go
|
|
||||||
// Add a mutex for active deploy operations
|
|
||||||
type Service struct {
|
|
||||||
// existing fields...
|
|
||||||
activeDeployMu sync.Mutex
|
|
||||||
}
|
|
||||||
|
|
||||||
// In Deploy function
|
|
||||||
func (svc *Service) Deploy(ctx context.Context, app *models.App, webhookEventID *int64, cancelExisting bool) error {
|
|
||||||
svc.activeDeployMu.Lock()
|
|
||||||
if cancelExisting {
|
|
||||||
svc.cancelActiveDeploy(app.ID)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Try to acquire per-app deployment lock
|
|
||||||
if !svc.tryLockApp(app.ID) {
|
|
||||||
svc.activeDeployMu.Unlock()
|
|
||||||
svc.log.Warn("deployment already in progress", "app", app.Name)
|
|
||||||
return ErrDeploymentInProgress
|
|
||||||
}
|
|
||||||
svc.activeDeployMu.Unlock()
|
|
||||||
|
|
||||||
defer svc.unlockApp(app.ID)
|
|
||||||
// rest of function...
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
## 6. Incomplete Error Propagation in Git Clone
|
|
||||||
|
|
||||||
### Description
|
|
||||||
In the Docker client's `runGitClone` function, if `ContainerLogs` fails, the error is silently ignored, which could hide important debugging information.
|
|
||||||
|
|
||||||
### Location
|
|
||||||
`internal/docker/client.go:679` in `runGitClone` function
|
|
||||||
|
|
||||||
### Proposed Fix
|
|
||||||
Handle the ContainerLogs error properly:
|
|
||||||
|
|
||||||
```go
|
|
||||||
// Always capture logs for the result
|
|
||||||
logs, logErr := c.ContainerLogs(ctx, containerID, "100")
|
|
||||||
if logErr != nil {
|
|
||||||
c.log.Warn("failed to get git clone logs", "error", logErr)
|
|
||||||
logs = "Failed to retrieve logs: " + logErr.Error()
|
|
||||||
}
|
|
||||||
```
|
|
||||||
68
CLAUDE.md
68
CLAUDE.md
@@ -1,68 +0,0 @@
|
|||||||
# Repository Rules
|
|
||||||
|
|
||||||
Last Updated 2026-01-08
|
|
||||||
|
|
||||||
These rules MUST be followed at all times, it is very important.
|
|
||||||
|
|
||||||
* Never use `git add -A` - add specific changes to a deliberate commit. A
|
|
||||||
commit should contain one change. After each change, make a commit with a
|
|
||||||
good one-line summary.
|
|
||||||
|
|
||||||
* NEVER modify the linter config without asking first.
|
|
||||||
|
|
||||||
* NEVER modify tests to exclude special cases or otherwise get them to pass
|
|
||||||
without asking first. In almost all cases, the code should be changed,
|
|
||||||
NOT the tests. If you think the test needs to be changed, make your case
|
|
||||||
for that and ask for permission to proceed, then stop. You need explicit
|
|
||||||
user approval to modify existing tests. (You do not need user approval
|
|
||||||
for writing NEW tests.)
|
|
||||||
|
|
||||||
* When linting, assume the linter config is CORRECT, and that each item
|
|
||||||
output by the linter is something that legitimately needs fixing in the
|
|
||||||
code.
|
|
||||||
|
|
||||||
* When running tests, use `make test`.
|
|
||||||
|
|
||||||
* Before commits, run `make check`. This runs `make lint` and `make test`
|
|
||||||
and `make check-fmt`. Any issues discovered MUST be resolved before
|
|
||||||
committing unless explicitly told otherwise.
|
|
||||||
|
|
||||||
* When fixing a bug, write a failing test for the bug FIRST. Add
|
|
||||||
appropriate logging to the test to ensure it is written correctly. Commit
|
|
||||||
that. Then go about fixing the bug until the test passes (without
|
|
||||||
modifying the test further). Then commit that.
|
|
||||||
|
|
||||||
* When adding a new feature, do the same - implement a test first (TDD). It
|
|
||||||
doesn't have to be super complex. Commit the test, then commit the
|
|
||||||
feature.
|
|
||||||
|
|
||||||
* When adding a new feature, use a feature branch. When the feature is
|
|
||||||
completely finished and the code is up to standards (passes `make check`)
|
|
||||||
then and only then can the feature branch be merged into `main` and the
|
|
||||||
branch deleted.
|
|
||||||
|
|
||||||
* Write godoc documentation comments for all exported types and functions as
|
|
||||||
you go along.
|
|
||||||
|
|
||||||
* ALWAYS be consistent in naming. If you name something one thing in one
|
|
||||||
place, name it the EXACT SAME THING in another place.
|
|
||||||
|
|
||||||
* Be descriptive and specific in naming. `wl` is bad;
|
|
||||||
`SourceHostWhitelist` is good. `ConnsPerHost` is bad;
|
|
||||||
`MaxConnectionsPerHost` is good.
|
|
||||||
|
|
||||||
* This is not prototype or teaching code - this is designed for production.
|
|
||||||
Any security issues (such as denial of service) or other web
|
|
||||||
vulnerabilities are P1 bugs and must be added to TODO.md at the top.
|
|
||||||
|
|
||||||
* As this is production code, no stubbing of implementations unless
|
|
||||||
specifically instructed. We need working implementations.
|
|
||||||
|
|
||||||
* Avoid vendoring deps unless specifically instructed to. NEVER commit
|
|
||||||
the vendor directory, NEVER commit compiled binaries. If these
|
|
||||||
directories or files exist, add them to .gitignore (and commit the
|
|
||||||
.gitignore) if they are not already in there. Keep the entire git
|
|
||||||
repository (with history) small - under 20MiB, unless you specifically
|
|
||||||
must commit larger files (e.g. test fixture example media files). Only
|
|
||||||
OUR source code and immediately supporting files (such as test examples)
|
|
||||||
goes into the repo/history.
|
|
||||||
@@ -51,7 +51,8 @@ type Config struct {
|
|||||||
MaintenanceMode bool
|
MaintenanceMode bool
|
||||||
MetricsUsername string
|
MetricsUsername string
|
||||||
MetricsPassword string
|
MetricsPassword string
|
||||||
SessionSecret string
|
SessionSecret string `json:"-"`
|
||||||
|
CORSOrigins string
|
||||||
params *Params
|
params *Params
|
||||||
log *slog.Logger
|
log *slog.Logger
|
||||||
}
|
}
|
||||||
@@ -102,6 +103,7 @@ func setupViper(name string) {
|
|||||||
viper.SetDefault("METRICS_USERNAME", "")
|
viper.SetDefault("METRICS_USERNAME", "")
|
||||||
viper.SetDefault("METRICS_PASSWORD", "")
|
viper.SetDefault("METRICS_PASSWORD", "")
|
||||||
viper.SetDefault("SESSION_SECRET", "")
|
viper.SetDefault("SESSION_SECRET", "")
|
||||||
|
viper.SetDefault("CORS_ORIGINS", "")
|
||||||
}
|
}
|
||||||
|
|
||||||
func buildConfig(log *slog.Logger, params *Params) (*Config, error) {
|
func buildConfig(log *slog.Logger, params *Params) (*Config, error) {
|
||||||
@@ -136,6 +138,7 @@ func buildConfig(log *slog.Logger, params *Params) (*Config, error) {
|
|||||||
MetricsUsername: viper.GetString("METRICS_USERNAME"),
|
MetricsUsername: viper.GetString("METRICS_USERNAME"),
|
||||||
MetricsPassword: viper.GetString("METRICS_PASSWORD"),
|
MetricsPassword: viper.GetString("METRICS_PASSWORD"),
|
||||||
SessionSecret: viper.GetString("SESSION_SECRET"),
|
SessionSecret: viper.GetString("SESSION_SECRET"),
|
||||||
|
CORSOrigins: viper.GetString("CORS_ORIGINS"),
|
||||||
params: params,
|
params: params,
|
||||||
log: log,
|
log: log,
|
||||||
}
|
}
|
||||||
@@ -154,10 +157,10 @@ func buildConfig(log *slog.Logger, params *Params) (*Config, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func loadOrCreateSessionSecret(log *slog.Logger, dataDir string) (string, error) {
|
func loadOrCreateSessionSecret(log *slog.Logger, dataDir string) (string, error) {
|
||||||
secretPath := filepath.Join(dataDir, sessionSecretFile)
|
secretPath := filepath.Clean(filepath.Join(dataDir, sessionSecretFile))
|
||||||
|
|
||||||
// Try to read existing secret
|
// Try to read existing secret
|
||||||
//nolint:gosec // secretPath is constructed from trusted config, not user input
|
// secretPath is constructed from trusted config (dataDir) and a constant filename.
|
||||||
data, err := os.ReadFile(secretPath)
|
data, err := os.ReadFile(secretPath)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
log.Info("loaded session secret from file", "path", secretPath)
|
log.Info("loaded session secret from file", "path", secretPath)
|
||||||
|
|||||||
@@ -1,6 +0,0 @@
|
|||||||
-- Initialize migrations table for tracking applied migrations
|
|
||||||
CREATE TABLE IF NOT EXISTS migrations (
|
|
||||||
id INTEGER PRIMARY KEY,
|
|
||||||
name TEXT NOT NULL UNIQUE,
|
|
||||||
applied_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
|
||||||
);
|
|
||||||
@@ -1,8 +1,7 @@
|
|||||||
-- Complete schema for upaas (consolidated)
|
-- Initial schema for upaas
|
||||||
-- This represents the final state of all migrations applied
|
|
||||||
|
|
||||||
-- Users table (single admin user)
|
-- Users table (single admin user)
|
||||||
CREATE TABLE IF NOT EXISTS users (
|
CREATE TABLE users (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
username TEXT UNIQUE NOT NULL,
|
username TEXT UNIQUE NOT NULL,
|
||||||
password_hash TEXT NOT NULL,
|
password_hash TEXT NOT NULL,
|
||||||
@@ -10,7 +9,7 @@ CREATE TABLE IF NOT EXISTS users (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- Apps table
|
-- Apps table
|
||||||
CREATE TABLE IF NOT EXISTS apps (
|
CREATE TABLE apps (
|
||||||
id TEXT PRIMARY KEY,
|
id TEXT PRIMARY KEY,
|
||||||
name TEXT UNIQUE NOT NULL,
|
name TEXT UNIQUE NOT NULL,
|
||||||
repo_url TEXT NOT NULL,
|
repo_url TEXT NOT NULL,
|
||||||
@@ -19,19 +18,18 @@ CREATE TABLE IF NOT EXISTS apps (
|
|||||||
webhook_secret TEXT NOT NULL,
|
webhook_secret TEXT NOT NULL,
|
||||||
ssh_private_key TEXT NOT NULL,
|
ssh_private_key TEXT NOT NULL,
|
||||||
ssh_public_key TEXT NOT NULL,
|
ssh_public_key TEXT NOT NULL,
|
||||||
|
container_id TEXT,
|
||||||
image_id TEXT,
|
image_id TEXT,
|
||||||
previous_image_id TEXT,
|
|
||||||
status TEXT DEFAULT 'pending',
|
status TEXT DEFAULT 'pending',
|
||||||
docker_network TEXT,
|
docker_network TEXT,
|
||||||
ntfy_topic TEXT,
|
ntfy_topic TEXT,
|
||||||
slack_webhook TEXT,
|
slack_webhook TEXT,
|
||||||
webhook_secret_hash TEXT NOT NULL DEFAULT '',
|
|
||||||
created_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
created_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
||||||
updated_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
updated_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
||||||
);
|
);
|
||||||
|
|
||||||
-- App environment variables
|
-- App environment variables
|
||||||
CREATE TABLE IF NOT EXISTS app_env_vars (
|
CREATE TABLE app_env_vars (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
key TEXT NOT NULL,
|
key TEXT NOT NULL,
|
||||||
@@ -40,7 +38,7 @@ CREATE TABLE IF NOT EXISTS app_env_vars (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- App labels
|
-- App labels
|
||||||
CREATE TABLE IF NOT EXISTS app_labels (
|
CREATE TABLE app_labels (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
key TEXT NOT NULL,
|
key TEXT NOT NULL,
|
||||||
@@ -49,7 +47,7 @@ CREATE TABLE IF NOT EXISTS app_labels (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- App volume mounts
|
-- App volume mounts
|
||||||
CREATE TABLE IF NOT EXISTS app_volumes (
|
CREATE TABLE app_volumes (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
host_path TEXT NOT NULL,
|
host_path TEXT NOT NULL,
|
||||||
@@ -57,24 +55,13 @@ CREATE TABLE IF NOT EXISTS app_volumes (
|
|||||||
readonly INTEGER DEFAULT 0
|
readonly INTEGER DEFAULT 0
|
||||||
);
|
);
|
||||||
|
|
||||||
-- App port mappings
|
|
||||||
CREATE TABLE IF NOT EXISTS app_ports (
|
|
||||||
id INTEGER PRIMARY KEY,
|
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
|
||||||
host_port INTEGER NOT NULL,
|
|
||||||
container_port INTEGER NOT NULL,
|
|
||||||
protocol TEXT NOT NULL DEFAULT 'tcp' CHECK(protocol IN ('tcp', 'udp')),
|
|
||||||
UNIQUE(host_port, protocol)
|
|
||||||
);
|
|
||||||
|
|
||||||
-- Webhook events log
|
-- Webhook events log
|
||||||
CREATE TABLE IF NOT EXISTS webhook_events (
|
CREATE TABLE webhook_events (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
event_type TEXT NOT NULL,
|
event_type TEXT NOT NULL,
|
||||||
branch TEXT NOT NULL,
|
branch TEXT NOT NULL,
|
||||||
commit_sha TEXT,
|
commit_sha TEXT,
|
||||||
commit_url TEXT,
|
|
||||||
payload TEXT,
|
payload TEXT,
|
||||||
matched INTEGER NOT NULL,
|
matched INTEGER NOT NULL,
|
||||||
processed INTEGER DEFAULT 0,
|
processed INTEGER DEFAULT 0,
|
||||||
@@ -82,13 +69,13 @@ CREATE TABLE IF NOT EXISTS webhook_events (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- Deployments log
|
-- Deployments log
|
||||||
CREATE TABLE IF NOT EXISTS deployments (
|
CREATE TABLE deployments (
|
||||||
id INTEGER PRIMARY KEY,
|
id INTEGER PRIMARY KEY,
|
||||||
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
webhook_event_id INTEGER REFERENCES webhook_events(id),
|
webhook_event_id INTEGER REFERENCES webhook_events(id),
|
||||||
commit_sha TEXT,
|
commit_sha TEXT,
|
||||||
commit_url TEXT,
|
|
||||||
image_id TEXT,
|
image_id TEXT,
|
||||||
|
container_id TEXT,
|
||||||
status TEXT NOT NULL,
|
status TEXT NOT NULL,
|
||||||
logs TEXT,
|
logs TEXT,
|
||||||
started_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
started_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
||||||
@@ -96,14 +83,12 @@ CREATE TABLE IF NOT EXISTS deployments (
|
|||||||
);
|
);
|
||||||
|
|
||||||
-- Indexes
|
-- Indexes
|
||||||
CREATE INDEX IF NOT EXISTS idx_apps_status ON apps(status);
|
CREATE INDEX idx_apps_status ON apps(status);
|
||||||
CREATE INDEX IF NOT EXISTS idx_apps_webhook_secret ON apps(webhook_secret);
|
CREATE INDEX idx_apps_webhook_secret ON apps(webhook_secret);
|
||||||
CREATE INDEX IF NOT EXISTS idx_apps_webhook_secret_hash ON apps(webhook_secret_hash);
|
CREATE INDEX idx_app_env_vars_app_id ON app_env_vars(app_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_app_env_vars_app_id ON app_env_vars(app_id);
|
CREATE INDEX idx_app_labels_app_id ON app_labels(app_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_app_labels_app_id ON app_labels(app_id);
|
CREATE INDEX idx_app_volumes_app_id ON app_volumes(app_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_app_volumes_app_id ON app_volumes(app_id);
|
CREATE INDEX idx_webhook_events_app_id ON webhook_events(app_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_app_ports_app_id ON app_ports(app_id);
|
CREATE INDEX idx_webhook_events_created_at ON webhook_events(created_at);
|
||||||
CREATE INDEX IF NOT EXISTS idx_webhook_events_app_id ON webhook_events(app_id);
|
CREATE INDEX idx_deployments_app_id ON deployments(app_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_webhook_events_created_at ON webhook_events(created_at);
|
CREATE INDEX idx_deployments_started_at ON deployments(started_at);
|
||||||
CREATE INDEX IF NOT EXISTS idx_deployments_app_id ON deployments(app_id);
|
|
||||||
CREATE INDEX IF NOT EXISTS idx_deployments_started_at ON deployments(started_at);
|
|
||||||
44
internal/database/migrations/002_remove_container_id.sql
Normal file
44
internal/database/migrations/002_remove_container_id.sql
Normal file
@@ -0,0 +1,44 @@
|
|||||||
|
-- Remove container_id from apps table
|
||||||
|
-- Container is now looked up via Docker label (upaas.id) instead of stored in database
|
||||||
|
|
||||||
|
-- SQLite doesn't support DROP COLUMN before version 3.35.0 (2021-03-12)
|
||||||
|
-- Use table rebuild for broader compatibility
|
||||||
|
|
||||||
|
-- Create new table without container_id
|
||||||
|
CREATE TABLE apps_new (
|
||||||
|
id TEXT PRIMARY KEY,
|
||||||
|
name TEXT UNIQUE NOT NULL,
|
||||||
|
repo_url TEXT NOT NULL,
|
||||||
|
branch TEXT NOT NULL DEFAULT 'main',
|
||||||
|
dockerfile_path TEXT DEFAULT 'Dockerfile',
|
||||||
|
webhook_secret TEXT NOT NULL,
|
||||||
|
ssh_private_key TEXT NOT NULL,
|
||||||
|
ssh_public_key TEXT NOT NULL,
|
||||||
|
image_id TEXT,
|
||||||
|
status TEXT DEFAULT 'pending',
|
||||||
|
docker_network TEXT,
|
||||||
|
ntfy_topic TEXT,
|
||||||
|
slack_webhook TEXT,
|
||||||
|
created_at DATETIME DEFAULT CURRENT_TIMESTAMP,
|
||||||
|
updated_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
-- Copy data (excluding container_id)
|
||||||
|
INSERT INTO apps_new (
|
||||||
|
id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
||||||
|
ssh_private_key, ssh_public_key, image_id, status,
|
||||||
|
docker_network, ntfy_topic, slack_webhook, created_at, updated_at
|
||||||
|
)
|
||||||
|
SELECT
|
||||||
|
id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
||||||
|
ssh_private_key, ssh_public_key, image_id, status,
|
||||||
|
docker_network, ntfy_topic, slack_webhook, created_at, updated_at
|
||||||
|
FROM apps;
|
||||||
|
|
||||||
|
-- Drop old table and rename new one
|
||||||
|
DROP TABLE apps;
|
||||||
|
ALTER TABLE apps_new RENAME TO apps;
|
||||||
|
|
||||||
|
-- Recreate indexes
|
||||||
|
CREATE INDEX idx_apps_status ON apps(status);
|
||||||
|
CREATE INDEX idx_apps_webhook_secret ON apps(webhook_secret);
|
||||||
12
internal/database/migrations/003_add_ports.sql
Normal file
12
internal/database/migrations/003_add_ports.sql
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
-- Add port mappings for apps
|
||||||
|
|
||||||
|
CREATE TABLE app_ports (
|
||||||
|
id INTEGER PRIMARY KEY,
|
||||||
|
app_id TEXT NOT NULL REFERENCES apps(id) ON DELETE CASCADE,
|
||||||
|
host_port INTEGER NOT NULL,
|
||||||
|
container_port INTEGER NOT NULL,
|
||||||
|
protocol TEXT NOT NULL DEFAULT 'tcp' CHECK(protocol IN ('tcp', 'udp')),
|
||||||
|
UNIQUE(host_port, protocol)
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE INDEX idx_app_ports_app_id ON app_ports(app_id);
|
||||||
3
internal/database/migrations/004_add_commit_url.sql
Normal file
3
internal/database/migrations/004_add_commit_url.sql
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
-- Add commit_url column to webhook_events and deployments tables
|
||||||
|
ALTER TABLE webhook_events ADD COLUMN commit_url TEXT;
|
||||||
|
ALTER TABLE deployments ADD COLUMN commit_url TEXT;
|
||||||
@@ -0,0 +1,2 @@
|
|||||||
|
-- Add webhook_secret_hash column for constant-time secret lookup
|
||||||
|
ALTER TABLE apps ADD COLUMN webhook_secret_hash TEXT NOT NULL DEFAULT '';
|
||||||
@@ -0,0 +1,2 @@
|
|||||||
|
-- Add previous_image_id to apps for deployment rollback support
|
||||||
|
ALTER TABLE apps ADD COLUMN previous_image_id TEXT;
|
||||||
@@ -17,6 +17,7 @@ import (
|
|||||||
"github.com/docker/docker/api/types"
|
"github.com/docker/docker/api/types"
|
||||||
"github.com/docker/docker/api/types/container"
|
"github.com/docker/docker/api/types/container"
|
||||||
"github.com/docker/docker/api/types/filters"
|
"github.com/docker/docker/api/types/filters"
|
||||||
|
"github.com/docker/docker/api/types/image"
|
||||||
"github.com/docker/docker/api/types/mount"
|
"github.com/docker/docker/api/types/mount"
|
||||||
"github.com/docker/docker/api/types/network"
|
"github.com/docker/docker/api/types/network"
|
||||||
"github.com/docker/docker/client"
|
"github.com/docker/docker/client"
|
||||||
@@ -479,6 +480,20 @@ func (c *Client) CloneRepo(
|
|||||||
return c.performClone(ctx, cfg)
|
return c.performClone(ctx, cfg)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// RemoveImage removes a Docker image by ID or tag.
|
||||||
|
// It returns nil if the image was successfully removed or does not exist.
|
||||||
|
func (c *Client) RemoveImage(ctx context.Context, imageID string) error {
|
||||||
|
_, err := c.docker.ImageRemove(ctx, imageID, image.RemoveOptions{
|
||||||
|
Force: true,
|
||||||
|
PruneChildren: true,
|
||||||
|
})
|
||||||
|
if err != nil && !client.IsErrNotFound(err) {
|
||||||
|
return fmt.Errorf("failed to remove image %s: %w", imageID, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func (c *Client) performBuild(
|
func (c *Client) performBuild(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
opts BuildImageOptions,
|
opts BuildImageOptions,
|
||||||
|
|||||||
@@ -74,18 +74,13 @@ func deploymentToAPI(d *models.Deployment) apiDeploymentResponse {
|
|||||||
// HandleAPILoginPOST returns a handler that authenticates via JSON credentials
|
// HandleAPILoginPOST returns a handler that authenticates via JSON credentials
|
||||||
// and sets a session cookie.
|
// and sets a session cookie.
|
||||||
func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
||||||
type loginRequest struct {
|
|
||||||
Username string `json:"username"`
|
|
||||||
Password string `json:"password"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type loginResponse struct {
|
type loginResponse struct {
|
||||||
UserID int64 `json:"userId"`
|
UserID int64 `json:"userId"`
|
||||||
Username string `json:"username"`
|
Username string `json:"username"`
|
||||||
}
|
}
|
||||||
|
|
||||||
return func(writer http.ResponseWriter, request *http.Request) {
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
var req loginRequest
|
var req map[string]string
|
||||||
|
|
||||||
decodeErr := json.NewDecoder(request.Body).Decode(&req)
|
decodeErr := json.NewDecoder(request.Body).Decode(&req)
|
||||||
if decodeErr != nil {
|
if decodeErr != nil {
|
||||||
@@ -96,7 +91,10 @@ func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
if req.Username == "" || req.Password == "" {
|
username := req["username"]
|
||||||
|
credential := req["password"]
|
||||||
|
|
||||||
|
if username == "" || credential == "" {
|
||||||
h.respondJSON(writer, request,
|
h.respondJSON(writer, request,
|
||||||
map[string]string{"error": "username and password are required"},
|
map[string]string{"error": "username and password are required"},
|
||||||
http.StatusBadRequest)
|
http.StatusBadRequest)
|
||||||
@@ -104,7 +102,7 @@ func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
user, authErr := h.auth.Authenticate(request.Context(), req.Username, req.Password)
|
user, authErr := h.auth.Authenticate(request.Context(), username, credential)
|
||||||
if authErr != nil {
|
if authErr != nil {
|
||||||
h.respondJSON(writer, request,
|
h.respondJSON(writer, request,
|
||||||
map[string]string{"error": "invalid credentials"},
|
map[string]string{"error": "invalid credentials"},
|
||||||
|
|||||||
@@ -6,6 +6,7 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
@@ -499,7 +500,11 @@ func (h *Handlers) HandleAppLogs() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
_, _ = writer.Write([]byte(logs))
|
// Container log output is attacker-controlled (untrusted) data.
|
||||||
|
// SanitizeLogs strips ANSI escapes and control characters.
|
||||||
|
// Content-Type is text/plain; XSS is not possible in this context.
|
||||||
|
sanitized := SanitizeLogs(logs)
|
||||||
|
_, _ = io.WriteString(writer, sanitized) // #nosec G705 -- text/plain Content-Type, SanitizeLogs strips control chars
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -534,7 +539,7 @@ func (h *Handlers) HandleDeploymentLogsAPI() http.HandlerFunc {
|
|||||||
|
|
||||||
logs := ""
|
logs := ""
|
||||||
if deployment.Logs.Valid {
|
if deployment.Logs.Valid {
|
||||||
logs = deployment.Logs.String
|
logs = SanitizeLogs(deployment.Logs.String)
|
||||||
}
|
}
|
||||||
|
|
||||||
response := map[string]any{
|
response := map[string]any{
|
||||||
@@ -581,8 +586,15 @@ func (h *Handlers) HandleDeploymentLogDownload() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if file exists
|
// Check if file exists — logPath is from GetLogFilePath (internal, not user input).
|
||||||
_, err := os.Stat(logPath)
|
// filepath.Clean normalizes the path and filepath.Base extracts the filename
|
||||||
|
// to prevent directory traversal.
|
||||||
|
cleanPath := filepath.Clean(logPath)
|
||||||
|
safeDir := filepath.Dir(cleanPath)
|
||||||
|
safeName := filepath.Base(cleanPath)
|
||||||
|
safePath := filepath.Join(safeDir, safeName)
|
||||||
|
|
||||||
|
_, err := os.Stat(safePath) // #nosec G703 -- path from internal GetLogFilePath, not user input
|
||||||
if os.IsNotExist(err) {
|
if os.IsNotExist(err) {
|
||||||
http.NotFound(writer, request)
|
http.NotFound(writer, request)
|
||||||
|
|
||||||
@@ -590,19 +602,19 @@ func (h *Handlers) HandleDeploymentLogDownload() http.HandlerFunc {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
h.log.Error("failed to stat log file", "error", err, "path", logPath)
|
h.log.Error("failed to stat log file", "error", err, "path", safePath)
|
||||||
http.Error(writer, "Internal Server Error", http.StatusInternalServerError)
|
http.Error(writer, "Internal Server Error", http.StatusInternalServerError)
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Extract filename for Content-Disposition header
|
// Extract filename for Content-Disposition header
|
||||||
filename := filepath.Base(logPath)
|
filename := safeName
|
||||||
|
|
||||||
writer.Header().Set("Content-Type", "text/plain; charset=utf-8")
|
writer.Header().Set("Content-Type", "text/plain; charset=utf-8")
|
||||||
writer.Header().Set("Content-Disposition", "attachment; filename=\""+filename+"\"")
|
writer.Header().Set("Content-Disposition", "attachment; filename=\""+filename+"\"")
|
||||||
|
|
||||||
http.ServeFile(writer, request, logPath)
|
http.ServeFile(writer, request, safePath)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -661,7 +673,7 @@ func (h *Handlers) HandleContainerLogsAPI() http.HandlerFunc {
|
|||||||
}
|
}
|
||||||
|
|
||||||
response := map[string]any{
|
response := map[string]any{
|
||||||
"logs": logs,
|
"logs": SanitizeLogs(logs),
|
||||||
"status": status,
|
"status": status,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
30
internal/handlers/sanitize.go
Normal file
30
internal/handlers/sanitize.go
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
package handlers
|
||||||
|
|
||||||
|
import (
|
||||||
|
"regexp"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ansiEscapePattern matches ANSI escape sequences (CSI, OSC, and single-character escapes).
|
||||||
|
var ansiEscapePattern = regexp.MustCompile(`(\x1b\[[0-9;]*[a-zA-Z]|\x1b\][^\x07]*\x07|\x1b[^[\]])`)
|
||||||
|
|
||||||
|
// SanitizeLogs strips ANSI escape sequences and non-printable control characters
|
||||||
|
// from container log output. Newlines (\n), carriage returns (\r), and tabs (\t)
|
||||||
|
// are preserved. This ensures that attacker-controlled container output cannot
|
||||||
|
// inject terminal escape sequences or other dangerous control characters.
|
||||||
|
func SanitizeLogs(input string) string {
|
||||||
|
// Strip ANSI escape sequences
|
||||||
|
result := ansiEscapePattern.ReplaceAllString(input, "")
|
||||||
|
|
||||||
|
// Strip remaining non-printable characters (keep \n, \r, \t)
|
||||||
|
var b strings.Builder
|
||||||
|
b.Grow(len(result))
|
||||||
|
|
||||||
|
for _, r := range result {
|
||||||
|
if r == '\n' || r == '\r' || r == '\t' || r >= ' ' {
|
||||||
|
b.WriteRune(r)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return b.String()
|
||||||
|
}
|
||||||
84
internal/handlers/sanitize_test.go
Normal file
84
internal/handlers/sanitize_test.go
Normal file
@@ -0,0 +1,84 @@
|
|||||||
|
package handlers_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/handlers"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestSanitizeLogs(t *testing.T) { //nolint:funlen // table-driven tests
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
input string
|
||||||
|
expected string
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "plain text unchanged",
|
||||||
|
input: "hello world\n",
|
||||||
|
expected: "hello world\n",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "strips ANSI color codes",
|
||||||
|
input: "\x1b[31mERROR\x1b[0m: something failed\n",
|
||||||
|
expected: "ERROR: something failed\n",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "strips OSC sequences",
|
||||||
|
input: "\x1b]0;window title\x07normal text\n",
|
||||||
|
expected: "normal text\n",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "strips null bytes",
|
||||||
|
input: "hello\x00world\n",
|
||||||
|
expected: "helloworld\n",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "strips bell characters",
|
||||||
|
input: "alert\x07here\n",
|
||||||
|
expected: "alerthere\n",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "preserves tabs",
|
||||||
|
input: "field1\tfield2\tfield3\n",
|
||||||
|
expected: "field1\tfield2\tfield3\n",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "preserves carriage returns",
|
||||||
|
input: "line1\r\nline2\r\n",
|
||||||
|
expected: "line1\r\nline2\r\n",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "strips mixed escape sequences",
|
||||||
|
input: "\x1b[32m2024-01-01\x1b[0m \x1b[1mINFO\x1b[0m starting\x00\n",
|
||||||
|
expected: "2024-01-01 INFO starting\n",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "empty string",
|
||||||
|
input: "",
|
||||||
|
expected: "",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "only control characters",
|
||||||
|
input: "\x00\x01\x02\x03",
|
||||||
|
expected: "",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "cursor movement sequences stripped",
|
||||||
|
input: "\x1b[2J\x1b[H\x1b[3Atext\n",
|
||||||
|
expected: "text\n",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
got := handlers.SanitizeLogs(tt.input)
|
||||||
|
if got != tt.expected {
|
||||||
|
t.Errorf("SanitizeLogs(%q) = %q, want %q", tt.input, got, tt.expected)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
83
internal/middleware/cors_test.go
Normal file
83
internal/middleware/cors_test.go
Normal file
@@ -0,0 +1,83 @@
|
|||||||
|
package middleware //nolint:testpackage // tests internal CORS behavior
|
||||||
|
|
||||||
|
import (
|
||||||
|
"log/slog"
|
||||||
|
"net/http"
|
||||||
|
"net/http/httptest"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
|
)
|
||||||
|
|
||||||
|
// testSessionValue is a dummy value for tests (not a real credential).
|
||||||
|
const testSessionValue = "test-value-32-bytes-long-enough!"
|
||||||
|
|
||||||
|
func newCORSTestMiddleware(corsOrigins string) *Middleware {
|
||||||
|
return &Middleware{
|
||||||
|
log: slog.Default(),
|
||||||
|
params: &Params{
|
||||||
|
Config: &config.Config{
|
||||||
|
CORSOrigins: corsOrigins,
|
||||||
|
SessionSecret: testSessionValue,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestCORS_NoOriginsConfigured_NoCORSHeaders(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
m := newCORSTestMiddleware("")
|
||||||
|
handler := m.CORS()(http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
|
||||||
|
w.WriteHeader(http.StatusOK)
|
||||||
|
}))
|
||||||
|
|
||||||
|
req := httptest.NewRequest(http.MethodGet, "/", nil)
|
||||||
|
req.Header.Set("Origin", "https://evil.com")
|
||||||
|
|
||||||
|
rec := httptest.NewRecorder()
|
||||||
|
handler.ServeHTTP(rec, req)
|
||||||
|
|
||||||
|
assert.Empty(t, rec.Header().Get("Access-Control-Allow-Origin"),
|
||||||
|
"expected no CORS headers when no origins configured")
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestCORS_OriginsConfigured_AllowsMatchingOrigin(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
m := newCORSTestMiddleware("https://app.example.com,https://other.example.com")
|
||||||
|
handler := m.CORS()(http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
|
||||||
|
w.WriteHeader(http.StatusOK)
|
||||||
|
}))
|
||||||
|
|
||||||
|
req := httptest.NewRequest(http.MethodGet, "/", nil)
|
||||||
|
req.Header.Set("Origin", "https://app.example.com")
|
||||||
|
|
||||||
|
rec := httptest.NewRecorder()
|
||||||
|
handler.ServeHTTP(rec, req)
|
||||||
|
|
||||||
|
assert.Equal(t, "https://app.example.com",
|
||||||
|
rec.Header().Get("Access-Control-Allow-Origin"))
|
||||||
|
assert.Equal(t, "true",
|
||||||
|
rec.Header().Get("Access-Control-Allow-Credentials"))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestCORS_OriginsConfigured_RejectsNonMatchingOrigin(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
m := newCORSTestMiddleware("https://app.example.com")
|
||||||
|
handler := m.CORS()(http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
|
||||||
|
w.WriteHeader(http.StatusOK)
|
||||||
|
}))
|
||||||
|
|
||||||
|
req := httptest.NewRequest(http.MethodGet, "/", nil)
|
||||||
|
req.Header.Set("Origin", "https://evil.com")
|
||||||
|
|
||||||
|
rec := httptest.NewRecorder()
|
||||||
|
handler.ServeHTTP(rec, req)
|
||||||
|
|
||||||
|
assert.Empty(t, rec.Header().Get("Access-Control-Allow-Origin"),
|
||||||
|
"expected no CORS headers for non-matching origin")
|
||||||
|
}
|
||||||
@@ -177,17 +177,48 @@ func realIP(r *http.Request) string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// CORS returns CORS middleware.
|
// CORS returns CORS middleware.
|
||||||
|
// When UPAAS_CORS_ORIGINS is empty (default), no CORS headers are sent
|
||||||
|
// (same-origin only). When configured, only the specified origins are
|
||||||
|
// allowed and credentials (cookies) are permitted.
|
||||||
func (m *Middleware) CORS() func(http.Handler) http.Handler {
|
func (m *Middleware) CORS() func(http.Handler) http.Handler {
|
||||||
|
origins := parseCORSOrigins(m.params.Config.CORSOrigins)
|
||||||
|
|
||||||
|
// No origins configured — no CORS headers (same-origin policy).
|
||||||
|
if len(origins) == 0 {
|
||||||
|
return func(next http.Handler) http.Handler {
|
||||||
|
return next
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
return cors.Handler(cors.Options{
|
return cors.Handler(cors.Options{
|
||||||
AllowedOrigins: []string{"*"},
|
AllowedOrigins: origins,
|
||||||
AllowedMethods: []string{"GET", "POST", "PUT", "DELETE", "OPTIONS"},
|
AllowedMethods: []string{"GET", "POST", "PUT", "DELETE", "OPTIONS"},
|
||||||
AllowedHeaders: []string{"Accept", "Authorization", "Content-Type", "X-CSRF-Token"},
|
AllowedHeaders: []string{"Accept", "Authorization", "Content-Type", "X-CSRF-Token"},
|
||||||
ExposedHeaders: []string{"Link"},
|
ExposedHeaders: []string{"Link"},
|
||||||
AllowCredentials: false,
|
AllowCredentials: true,
|
||||||
MaxAge: corsMaxAge,
|
MaxAge: corsMaxAge,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// parseCORSOrigins splits a comma-separated origin string into a slice,
|
||||||
|
// trimming whitespace. Returns nil if the input is empty.
|
||||||
|
func parseCORSOrigins(raw string) []string {
|
||||||
|
if raw == "" {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
parts := strings.Split(raw, ",")
|
||||||
|
origins := make([]string, 0, len(parts))
|
||||||
|
|
||||||
|
for _, p := range parts {
|
||||||
|
if o := strings.TrimSpace(p); o != "" {
|
||||||
|
origins = append(origins, o)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return origins
|
||||||
|
}
|
||||||
|
|
||||||
// MetricsAuth returns basic auth middleware for metrics endpoint.
|
// MetricsAuth returns basic auth middleware for metrics endpoint.
|
||||||
func (m *Middleware) MetricsAuth() func(http.Handler) http.Handler {
|
func (m *Middleware) MetricsAuth() func(http.Handler) http.Handler {
|
||||||
if m.params.Config.MetricsUsername == "" {
|
if m.params.Config.MetricsUsername == "" {
|
||||||
|
|||||||
@@ -11,6 +11,7 @@ import (
|
|||||||
"log/slog"
|
"log/slog"
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
@@ -472,7 +473,7 @@ func (svc *Service) runBuildAndDeploy(
|
|||||||
// Build phase with timeout
|
// Build phase with timeout
|
||||||
imageID, err := svc.buildImageWithTimeout(deployCtx, app, deployment)
|
imageID, err := svc.buildImageWithTimeout(deployCtx, app, deployment)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment)
|
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment, "")
|
||||||
if cancelErr != nil {
|
if cancelErr != nil {
|
||||||
return cancelErr
|
return cancelErr
|
||||||
}
|
}
|
||||||
@@ -485,7 +486,7 @@ func (svc *Service) runBuildAndDeploy(
|
|||||||
// Deploy phase with timeout
|
// Deploy phase with timeout
|
||||||
err = svc.deployContainerWithTimeout(deployCtx, app, deployment, imageID)
|
err = svc.deployContainerWithTimeout(deployCtx, app, deployment, imageID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment)
|
cancelErr := svc.checkCancelled(deployCtx, bgCtx, app, deployment, imageID)
|
||||||
if cancelErr != nil {
|
if cancelErr != nil {
|
||||||
return cancelErr
|
return cancelErr
|
||||||
}
|
}
|
||||||
@@ -661,24 +662,77 @@ func (svc *Service) cancelActiveDeploy(appID string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// checkCancelled checks if the deploy context was cancelled (by a newer deploy)
|
// checkCancelled checks if the deploy context was cancelled (by a newer deploy)
|
||||||
// and if so, marks the deployment as cancelled. Returns ErrDeployCancelled or nil.
|
// and if so, marks the deployment as cancelled and cleans up orphan resources.
|
||||||
|
// Returns ErrDeployCancelled or nil.
|
||||||
func (svc *Service) checkCancelled(
|
func (svc *Service) checkCancelled(
|
||||||
deployCtx context.Context,
|
deployCtx context.Context,
|
||||||
bgCtx context.Context,
|
bgCtx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
|
imageID string,
|
||||||
) error {
|
) error {
|
||||||
if !errors.Is(deployCtx.Err(), context.Canceled) {
|
if !errors.Is(deployCtx.Err(), context.Canceled) {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
svc.log.Info("deployment cancelled by newer deploy", "app", app.Name)
|
svc.log.Info("deployment cancelled", "app", app.Name)
|
||||||
|
|
||||||
|
svc.cleanupCancelledDeploy(bgCtx, app, deployment, imageID)
|
||||||
|
|
||||||
_ = deployment.MarkFinished(bgCtx, models.DeploymentStatusCancelled)
|
_ = deployment.MarkFinished(bgCtx, models.DeploymentStatusCancelled)
|
||||||
|
|
||||||
return ErrDeployCancelled
|
return ErrDeployCancelled
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// cleanupCancelledDeploy removes orphan resources left by a cancelled deployment.
|
||||||
|
func (svc *Service) cleanupCancelledDeploy(
|
||||||
|
ctx context.Context,
|
||||||
|
app *models.App,
|
||||||
|
deployment *models.Deployment,
|
||||||
|
imageID string,
|
||||||
|
) {
|
||||||
|
// Clean up the intermediate Docker image if one was built
|
||||||
|
if imageID != "" {
|
||||||
|
removeErr := svc.docker.RemoveImage(ctx, imageID)
|
||||||
|
if removeErr != nil {
|
||||||
|
svc.log.Error("failed to remove image from cancelled deploy",
|
||||||
|
"error", removeErr, "app", app.Name, "image", imageID)
|
||||||
|
_ = deployment.AppendLog(ctx, "WARNING: failed to clean up image "+imageID+": "+removeErr.Error())
|
||||||
|
} else {
|
||||||
|
svc.log.Info("cleaned up image from cancelled deploy",
|
||||||
|
"app", app.Name, "image", imageID)
|
||||||
|
_ = deployment.AppendLog(ctx, "Cleaned up intermediate image: "+imageID)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Clean up the build directory for this deployment
|
||||||
|
buildDir := svc.GetBuildDir(app.Name)
|
||||||
|
|
||||||
|
entries, err := os.ReadDir(buildDir)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
prefix := fmt.Sprintf("%d-", deployment.ID)
|
||||||
|
|
||||||
|
for _, entry := range entries {
|
||||||
|
if entry.IsDir() && strings.HasPrefix(entry.Name(), prefix) {
|
||||||
|
dirPath := filepath.Join(buildDir, entry.Name())
|
||||||
|
|
||||||
|
removeErr := os.RemoveAll(dirPath)
|
||||||
|
if removeErr != nil {
|
||||||
|
svc.log.Error("failed to remove build dir from cancelled deploy",
|
||||||
|
"error", removeErr, "path", dirPath)
|
||||||
|
} else {
|
||||||
|
svc.log.Info("cleaned up build dir from cancelled deploy",
|
||||||
|
"app", app.Name, "path", dirPath)
|
||||||
|
|
||||||
|
_ = deployment.AppendLog(ctx, "Cleaned up build directory")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func (svc *Service) fetchWebhookEvent(
|
func (svc *Service) fetchWebhookEvent(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
webhookEventID *int64,
|
webhookEventID *int64,
|
||||||
|
|||||||
63
internal/service/deploy/deploy_cleanup_test.go
Normal file
63
internal/service/deploy/deploy_cleanup_test.go
Normal file
@@ -0,0 +1,63 @@
|
|||||||
|
package deploy_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"log/slog"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestCleanupCancelledDeploy_RemovesBuildDir(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tmpDir := t.TempDir()
|
||||||
|
cfg := &config.Config{DataDir: tmpDir}
|
||||||
|
|
||||||
|
svc := deploy.NewTestServiceWithConfig(slog.Default(), cfg, nil)
|
||||||
|
|
||||||
|
// Create a fake build directory matching the deployment pattern
|
||||||
|
appName := "test-app"
|
||||||
|
buildDir := svc.GetBuildDirExported(appName)
|
||||||
|
require.NoError(t, os.MkdirAll(buildDir, 0o750))
|
||||||
|
|
||||||
|
// Create deployment-specific dir: <deploymentID>-<random>
|
||||||
|
deployDir := filepath.Join(buildDir, "42-abc123")
|
||||||
|
require.NoError(t, os.MkdirAll(deployDir, 0o750))
|
||||||
|
|
||||||
|
// Create a file inside to verify full removal
|
||||||
|
require.NoError(t, os.WriteFile(filepath.Join(deployDir, "work"), []byte("test"), 0o600))
|
||||||
|
|
||||||
|
// Also create a dir for a different deployment (should NOT be removed)
|
||||||
|
otherDir := filepath.Join(buildDir, "99-xyz789")
|
||||||
|
require.NoError(t, os.MkdirAll(otherDir, 0o750))
|
||||||
|
|
||||||
|
// Run cleanup for deployment 42
|
||||||
|
svc.CleanupCancelledDeploy(context.Background(), appName, 42, "")
|
||||||
|
|
||||||
|
// Deployment 42's dir should be gone
|
||||||
|
_, err := os.Stat(deployDir)
|
||||||
|
assert.True(t, os.IsNotExist(err), "deployment build dir should be removed")
|
||||||
|
|
||||||
|
// Deployment 99's dir should still exist
|
||||||
|
_, err = os.Stat(otherDir)
|
||||||
|
assert.NoError(t, err, "other deployment build dir should not be removed")
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestCleanupCancelledDeploy_NoBuildDir(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tmpDir := t.TempDir()
|
||||||
|
cfg := &config.Config{DataDir: tmpDir}
|
||||||
|
|
||||||
|
svc := deploy.NewTestServiceWithConfig(slog.Default(), cfg, nil)
|
||||||
|
|
||||||
|
// Should not panic when build dir doesn't exist
|
||||||
|
svc.CleanupCancelledDeploy(context.Background(), "nonexistent-app", 1, "")
|
||||||
|
}
|
||||||
@@ -2,7 +2,14 @@ package deploy
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"fmt"
|
||||||
"log/slog"
|
"log/slog"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/docker"
|
||||||
)
|
)
|
||||||
|
|
||||||
// NewTestService creates a Service with minimal dependencies for testing.
|
// NewTestService creates a Service with minimal dependencies for testing.
|
||||||
@@ -31,3 +38,45 @@ func (svc *Service) TryLockApp(appID string) bool {
|
|||||||
func (svc *Service) UnlockApp(appID string) {
|
func (svc *Service) UnlockApp(appID string) {
|
||||||
svc.unlockApp(appID)
|
svc.unlockApp(appID)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NewTestServiceWithConfig creates a Service with config and docker client for testing.
|
||||||
|
func NewTestServiceWithConfig(log *slog.Logger, cfg *config.Config, dockerClient *docker.Client) *Service {
|
||||||
|
return &Service{
|
||||||
|
log: log,
|
||||||
|
config: cfg,
|
||||||
|
docker: dockerClient,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// CleanupCancelledDeploy exposes the build directory cleanup portion of
|
||||||
|
// cleanupCancelledDeploy for testing. It removes build directories matching
|
||||||
|
// the deployment ID prefix.
|
||||||
|
func (svc *Service) CleanupCancelledDeploy(
|
||||||
|
_ context.Context,
|
||||||
|
appName string,
|
||||||
|
deploymentID int64,
|
||||||
|
_ string,
|
||||||
|
) {
|
||||||
|
// We can't create real models.App/Deployment in tests easily,
|
||||||
|
// so we test the build dir cleanup portion directly.
|
||||||
|
buildDir := svc.GetBuildDir(appName)
|
||||||
|
|
||||||
|
entries, err := os.ReadDir(buildDir)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
prefix := fmt.Sprintf("%d-", deploymentID)
|
||||||
|
|
||||||
|
for _, entry := range entries {
|
||||||
|
if entry.IsDir() && strings.HasPrefix(entry.Name(), prefix) {
|
||||||
|
dirPath := filepath.Join(buildDir, entry.Name())
|
||||||
|
_ = os.RemoveAll(dirPath)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetBuildDirExported exposes GetBuildDir for testing.
|
||||||
|
func (svc *Service) GetBuildDirExported(appName string) string {
|
||||||
|
return svc.GetBuildDir(appName)
|
||||||
|
}
|
||||||
|
|||||||
@@ -10,6 +10,7 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"net/url"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
@@ -241,12 +242,34 @@ func (svc *Service) sendNotifications(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// errInvalidURLScheme indicates the webhook URL uses a disallowed scheme.
|
||||||
|
var errInvalidURLScheme = errors.New("URL scheme not allowed, must be http or https")
|
||||||
|
|
||||||
|
// validateWebhookURL validates that a webhook URL is well-formed and uses http/https.
|
||||||
|
func validateWebhookURL(rawURL string) error {
|
||||||
|
parsed, err := url.ParseRequestURI(rawURL)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("malformed URL: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if parsed.Scheme != "https" && parsed.Scheme != "http" {
|
||||||
|
return fmt.Errorf("%w: got %q", errInvalidURLScheme, parsed.Scheme)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func (svc *Service) sendNtfy(
|
func (svc *Service) sendNtfy(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
topic, title, message, priority string,
|
topic, title, message, priority string,
|
||||||
) error {
|
) error {
|
||||||
svc.log.Debug("sending ntfy notification", "topic", topic, "title", title)
|
svc.log.Debug("sending ntfy notification", "topic", topic, "title", title)
|
||||||
|
|
||||||
|
urlErr := validateWebhookURL(topic)
|
||||||
|
if urlErr != nil {
|
||||||
|
return fmt.Errorf("invalid ntfy topic URL: %w", urlErr)
|
||||||
|
}
|
||||||
|
|
||||||
request, err := http.NewRequestWithContext(
|
request, err := http.NewRequestWithContext(
|
||||||
ctx,
|
ctx,
|
||||||
http.MethodPost,
|
http.MethodPost,
|
||||||
@@ -260,7 +283,7 @@ func (svc *Service) sendNtfy(
|
|||||||
request.Header.Set("Title", title)
|
request.Header.Set("Title", title)
|
||||||
request.Header.Set("Priority", svc.ntfyPriority(priority))
|
request.Header.Set("Priority", svc.ntfyPriority(priority))
|
||||||
|
|
||||||
resp, err := svc.client.Do(request)
|
resp, err := svc.client.Do(request) // #nosec G704 -- URL from validated config, not user input
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to send ntfy request: %w", err)
|
return fmt.Errorf("failed to send ntfy request: %w", err)
|
||||||
}
|
}
|
||||||
@@ -340,6 +363,11 @@ func (svc *Service) sendSlack(
|
|||||||
return fmt.Errorf("failed to marshal slack payload: %w", err)
|
return fmt.Errorf("failed to marshal slack payload: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
urlErr := validateWebhookURL(webhookURL)
|
||||||
|
if urlErr != nil {
|
||||||
|
return fmt.Errorf("invalid slack webhook URL: %w", urlErr)
|
||||||
|
}
|
||||||
|
|
||||||
request, err := http.NewRequestWithContext(
|
request, err := http.NewRequestWithContext(
|
||||||
ctx,
|
ctx,
|
||||||
http.MethodPost,
|
http.MethodPost,
|
||||||
@@ -352,7 +380,7 @@ func (svc *Service) sendSlack(
|
|||||||
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
request.Header.Set("Content-Type", "application/json")
|
||||||
|
|
||||||
resp, err := svc.client.Do(request)
|
resp, err := svc.client.Do(request) // #nosec G704 -- URL from validated config, not user input
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to send slack request: %w", err)
|
return fmt.Errorf("failed to send slack request: %w", err)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ import (
|
|||||||
|
|
||||||
// KeyPair contains an SSH key pair.
|
// KeyPair contains an SSH key pair.
|
||||||
type KeyPair struct {
|
type KeyPair struct {
|
||||||
PrivateKey string
|
PrivateKey string `json:"-"`
|
||||||
PublicKey string
|
PublicKey string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -369,7 +369,7 @@ document.addEventListener("alpine:init", () => {
|
|||||||
init() {
|
init() {
|
||||||
// Read initial logs from script tag (avoids escaping issues)
|
// Read initial logs from script tag (avoids escaping issues)
|
||||||
const initialLogsEl = this.$el.querySelector(".initial-logs");
|
const initialLogsEl = this.$el.querySelector(".initial-logs");
|
||||||
this.logs = initialLogsEl?.textContent || "Loading...";
|
this.logs = initialLogsEl?.dataset.logs || "Loading...";
|
||||||
|
|
||||||
// Set up scroll tracking
|
// Set up scroll tracking
|
||||||
this.$nextTick(() => {
|
this.$nextTick(() => {
|
||||||
|
|||||||
@@ -98,7 +98,7 @@
|
|||||||
title="Scroll to bottom"
|
title="Scroll to bottom"
|
||||||
>↓ Follow</button>
|
>↓ Follow</button>
|
||||||
</div>
|
</div>
|
||||||
{{if .Logs.Valid}}<script type="text/plain" class="initial-logs">{{.Logs.String}}</script>{{end}}
|
{{if .Logs.Valid}}<div hidden class="initial-logs" data-logs="{{.Logs.String}}"></div>{{end}}
|
||||||
</div>
|
</div>
|
||||||
{{end}}
|
{{end}}
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
Reference in New Issue
Block a user