Compare commits
1 Commits
feature/ba
...
feature/cp
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
d6f6cc3670 |
@@ -9,9 +9,9 @@ A simple self-hosted PaaS that auto-deploys Docker containers from Git repositor
|
||||
- Per-app UUID-based webhook URLs for Gitea integration
|
||||
- Branch filtering - only deploy on configured branch changes
|
||||
- Environment variables, labels, and volume mounts per app
|
||||
- CPU and memory resource limits per app
|
||||
- Docker builds via socket access
|
||||
- Notifications via ntfy and Slack-compatible webhooks
|
||||
- Backup/restore of app configurations (JSON export/import via UI and API)
|
||||
- Simple server-rendered UI with Tailwind CSS
|
||||
|
||||
## Non-Goals
|
||||
|
||||
3
internal/database/migrations/007_add_resource_limits.sql
Normal file
3
internal/database/migrations/007_add_resource_limits.sql
Normal file
@@ -0,0 +1,3 @@
|
||||
-- Add CPU and memory resource limits per app
|
||||
ALTER TABLE apps ADD COLUMN cpu_limit REAL;
|
||||
ALTER TABLE apps ADD COLUMN memory_limit INTEGER;
|
||||
@@ -138,13 +138,15 @@ func (c *Client) BuildImage(
|
||||
|
||||
// CreateContainerOptions contains options for creating a container.
|
||||
type CreateContainerOptions struct {
|
||||
Name string
|
||||
Image string
|
||||
Env map[string]string
|
||||
Labels map[string]string
|
||||
Volumes []VolumeMount
|
||||
Ports []PortMapping
|
||||
Network string
|
||||
Name string
|
||||
Image string
|
||||
Env map[string]string
|
||||
Labels map[string]string
|
||||
Volumes []VolumeMount
|
||||
Ports []PortMapping
|
||||
Network string
|
||||
CPULimit float64 // CPU cores (e.g. 0.5 = half a core, 2.0 = two cores). 0 means unlimited.
|
||||
MemoryLimit int64 // Memory in bytes. 0 means unlimited.
|
||||
}
|
||||
|
||||
// VolumeMount represents a volume mount.
|
||||
@@ -161,6 +163,14 @@ type PortMapping struct {
|
||||
Protocol string // "tcp" or "udp"
|
||||
}
|
||||
|
||||
// nanoCPUsPerCPU is the number of NanoCPUs per CPU core.
|
||||
const nanoCPUsPerCPU = 1e9
|
||||
|
||||
// cpuLimitToNanoCPUs converts a CPU limit (e.g. 0.5 cores) to Docker NanoCPUs.
|
||||
func cpuLimitToNanoCPUs(cpuLimit float64) int64 {
|
||||
return int64(cpuLimit * nanoCPUsPerCPU)
|
||||
}
|
||||
|
||||
// buildPortConfig converts port mappings to Docker port configuration.
|
||||
func buildPortConfig(ports []PortMapping) (nat.PortSet, nat.PortMap) {
|
||||
exposedPorts := make(nat.PortSet)
|
||||
@@ -185,6 +195,48 @@ func buildPortConfig(ports []PortMapping) (nat.PortSet, nat.PortMap) {
|
||||
return exposedPorts, portBindings
|
||||
}
|
||||
|
||||
// buildEnvSlice converts an env map to a Docker-compatible env slice.
|
||||
func buildEnvSlice(env map[string]string) []string {
|
||||
envSlice := make([]string, 0, len(env))
|
||||
|
||||
for key, val := range env {
|
||||
envSlice = append(envSlice, key+"="+val)
|
||||
}
|
||||
|
||||
return envSlice
|
||||
}
|
||||
|
||||
// buildMounts converts volume mounts to Docker mount configuration.
|
||||
func buildMounts(volumes []VolumeMount) []mount.Mount {
|
||||
mounts := make([]mount.Mount, 0, len(volumes))
|
||||
|
||||
for _, vol := range volumes {
|
||||
mounts = append(mounts, mount.Mount{
|
||||
Type: mount.TypeBind,
|
||||
Source: vol.HostPath,
|
||||
Target: vol.ContainerPath,
|
||||
ReadOnly: vol.ReadOnly,
|
||||
})
|
||||
}
|
||||
|
||||
return mounts
|
||||
}
|
||||
|
||||
// buildResources builds Docker resource constraints from container options.
|
||||
func buildResources(opts CreateContainerOptions) container.Resources {
|
||||
resources := container.Resources{}
|
||||
|
||||
if opts.CPULimit > 0 {
|
||||
resources.NanoCPUs = cpuLimitToNanoCPUs(opts.CPULimit)
|
||||
}
|
||||
|
||||
if opts.MemoryLimit > 0 {
|
||||
resources.Memory = opts.MemoryLimit
|
||||
}
|
||||
|
||||
return resources
|
||||
}
|
||||
|
||||
// CreateContainer creates a new container.
|
||||
func (c *Client) CreateContainer(
|
||||
ctx context.Context,
|
||||
@@ -196,40 +248,20 @@ func (c *Client) CreateContainer(
|
||||
|
||||
c.log.Info("creating container", "name", opts.Name, "image", opts.Image)
|
||||
|
||||
// Convert env map to slice
|
||||
envSlice := make([]string, 0, len(opts.Env))
|
||||
|
||||
for key, val := range opts.Env {
|
||||
envSlice = append(envSlice, key+"="+val)
|
||||
}
|
||||
|
||||
// Convert volumes to mounts
|
||||
mounts := make([]mount.Mount, 0, len(opts.Volumes))
|
||||
|
||||
for _, vol := range opts.Volumes {
|
||||
mounts = append(mounts, mount.Mount{
|
||||
Type: mount.TypeBind,
|
||||
Source: vol.HostPath,
|
||||
Target: vol.ContainerPath,
|
||||
ReadOnly: vol.ReadOnly,
|
||||
})
|
||||
}
|
||||
|
||||
// Convert ports to exposed ports and port bindings
|
||||
exposedPorts, portBindings := buildPortConfig(opts.Ports)
|
||||
|
||||
// Create container
|
||||
resp, err := c.docker.ContainerCreate(ctx,
|
||||
&container.Config{
|
||||
Image: opts.Image,
|
||||
Env: envSlice,
|
||||
Env: buildEnvSlice(opts.Env),
|
||||
Labels: opts.Labels,
|
||||
ExposedPorts: exposedPorts,
|
||||
},
|
||||
&container.HostConfig{
|
||||
Mounts: mounts,
|
||||
Mounts: buildMounts(opts.Volumes),
|
||||
PortBindings: portBindings,
|
||||
NetworkMode: container.NetworkMode(opts.Network),
|
||||
Resources: buildResources(opts),
|
||||
RestartPolicy: container.RestartPolicy{
|
||||
Name: container.RestartPolicyUnlessStopped,
|
||||
},
|
||||
|
||||
31
internal/docker/resource_limits_test.go
Normal file
31
internal/docker/resource_limits_test.go
Normal file
@@ -0,0 +1,31 @@
|
||||
package docker //nolint:testpackage // tests unexported cpuLimitToNanoCPUs
|
||||
|
||||
import (
|
||||
"testing"
|
||||
)
|
||||
|
||||
func TestCpuLimitToNanoCPUs(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
tests := []struct {
|
||||
name string
|
||||
cpuLimit float64
|
||||
expected int64
|
||||
}{
|
||||
{"one core", 1.0, 1_000_000_000},
|
||||
{"half core", 0.5, 500_000_000},
|
||||
{"two cores", 2.0, 2_000_000_000},
|
||||
{"quarter core", 0.25, 250_000_000},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
got := cpuLimitToNanoCPUs(tt.cpuLimit)
|
||||
if got != tt.expected {
|
||||
t.Errorf("cpuLimitToNanoCPUs(%v) = %d, want %d", tt.cpuLimit, got, tt.expected)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -27,11 +27,6 @@ func apiRouter(tc *testContext) http.Handler {
|
||||
apiR.Get("/apps", tc.handlers.HandleAPIListApps())
|
||||
apiR.Get("/apps/{id}", tc.handlers.HandleAPIGetApp())
|
||||
apiR.Get("/apps/{id}/deployments", tc.handlers.HandleAPIListDeployments())
|
||||
|
||||
// Backup/Restore API
|
||||
apiR.Get("/apps/{id}/export", tc.handlers.HandleAPIExportApp())
|
||||
apiR.Get("/backup/export", tc.handlers.HandleAPIExportAllApps())
|
||||
apiR.Post("/backup/import", tc.handlers.HandleAPIImportApps())
|
||||
})
|
||||
})
|
||||
|
||||
|
||||
@@ -257,23 +257,19 @@ func (h *Handlers) HandleAppUpdate() http.HandlerFunc { //nolint:funlen // valid
|
||||
application.RepoURL = request.FormValue("repo_url")
|
||||
application.Branch = request.FormValue("branch")
|
||||
application.DockerfilePath = request.FormValue("dockerfile_path")
|
||||
application.DockerNetwork = optionalNullString(request.FormValue("docker_network"))
|
||||
application.NtfyTopic = optionalNullString(request.FormValue("ntfy_topic"))
|
||||
application.SlackWebhook = optionalNullString(request.FormValue("slack_webhook"))
|
||||
|
||||
if network := request.FormValue("docker_network"); network != "" {
|
||||
application.DockerNetwork = sql.NullString{String: network, Valid: true}
|
||||
} else {
|
||||
application.DockerNetwork = sql.NullString{}
|
||||
}
|
||||
limitsErr := applyResourceLimits(application, request)
|
||||
if limitsErr != "" {
|
||||
data := h.addGlobals(map[string]any{
|
||||
"App": application,
|
||||
"Error": limitsErr,
|
||||
}, request)
|
||||
h.renderTemplate(writer, tmpl, "app_edit.html", data)
|
||||
|
||||
if ntfy := request.FormValue("ntfy_topic"); ntfy != "" {
|
||||
application.NtfyTopic = sql.NullString{String: ntfy, Valid: true}
|
||||
} else {
|
||||
application.NtfyTopic = sql.NullString{}
|
||||
}
|
||||
|
||||
if slack := request.FormValue("slack_webhook"); slack != "" {
|
||||
application.SlackWebhook = sql.NullString{String: slack, Valid: true}
|
||||
} else {
|
||||
application.SlackWebhook = sql.NullString{}
|
||||
return
|
||||
}
|
||||
|
||||
saveErr := application.Save(request.Context())
|
||||
@@ -1368,6 +1364,129 @@ func validateVolumePaths(hostPath, containerPath string) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// ErrInvalidMemoryFormat is returned when a memory limit string cannot be parsed.
|
||||
var ErrInvalidMemoryFormat = errors.New(
|
||||
"must be a number with optional unit suffix (e.g. 256m, 1g, 512000000)",
|
||||
)
|
||||
|
||||
// ErrNegativeValue is returned when a resource limit is negative.
|
||||
var ErrNegativeValue = errors.New("value must be positive")
|
||||
|
||||
// Memory unit byte multipliers.
|
||||
const (
|
||||
kilobyte = 1024
|
||||
megabyte = 1024 * 1024
|
||||
gigabyte = 1024 * 1024 * 1024
|
||||
)
|
||||
|
||||
// optionalNullString converts a form value to a sql.NullString.
|
||||
// Returns a valid NullString if non-empty, invalid (NULL) if empty.
|
||||
func optionalNullString(s string) sql.NullString {
|
||||
if s != "" {
|
||||
return sql.NullString{String: s, Valid: true}
|
||||
}
|
||||
|
||||
return sql.NullString{}
|
||||
}
|
||||
|
||||
// applyResourceLimits parses CPU and memory limit form values and applies them to the app.
|
||||
// Returns an error message string if validation fails, or empty string on success.
|
||||
func applyResourceLimits(application *models.App, request *http.Request) string {
|
||||
cpuLimit, cpuErr := parseOptionalFloat64(request.FormValue("cpu_limit"))
|
||||
if cpuErr != nil {
|
||||
return "Invalid CPU limit: must be a positive number (e.g. 0.5, 1, 2)"
|
||||
}
|
||||
|
||||
application.CPULimit = cpuLimit
|
||||
|
||||
memoryLimit, memErr := parseOptionalMemoryBytes(request.FormValue("memory_limit"))
|
||||
if memErr != nil {
|
||||
return "Invalid memory limit: " + memErr.Error()
|
||||
}
|
||||
|
||||
application.MemoryLimit = memoryLimit
|
||||
|
||||
return ""
|
||||
}
|
||||
|
||||
// memoryUnitMultiplier returns the byte multiplier for a memory unit suffix.
|
||||
// Returns 0 if the suffix is not recognized.
|
||||
func memoryUnitMultiplier(suffix byte) int64 {
|
||||
switch suffix {
|
||||
case 'k':
|
||||
return kilobyte
|
||||
case 'm':
|
||||
return megabyte
|
||||
case 'g':
|
||||
return gigabyte
|
||||
default:
|
||||
return 0
|
||||
}
|
||||
}
|
||||
|
||||
// parseOptionalFloat64 parses an optional float64 form field.
|
||||
// Returns a valid NullFloat64 if the string is non-empty and parses to a positive number.
|
||||
// Returns an empty NullFloat64 if the string is empty.
|
||||
// Returns an error if the string is non-empty but invalid or non-positive.
|
||||
func parseOptionalFloat64(s string) (sql.NullFloat64, error) {
|
||||
s = strings.TrimSpace(s)
|
||||
if s == "" {
|
||||
return sql.NullFloat64{}, nil
|
||||
}
|
||||
|
||||
val, err := strconv.ParseFloat(s, 64)
|
||||
if err != nil {
|
||||
return sql.NullFloat64{}, fmt.Errorf("invalid number: %w", err)
|
||||
}
|
||||
|
||||
if val <= 0 {
|
||||
return sql.NullFloat64{}, ErrNegativeValue
|
||||
}
|
||||
|
||||
return sql.NullFloat64{Float64: val, Valid: true}, nil
|
||||
}
|
||||
|
||||
// parseOptionalMemoryBytes parses an optional memory limit string into bytes.
|
||||
// Accepts plain bytes (e.g. "536870912") or suffixed values (e.g. "512m", "1g", "256k").
|
||||
// Returns a valid NullInt64 with bytes if non-empty, empty NullInt64 if blank.
|
||||
func parseOptionalMemoryBytes(s string) (sql.NullInt64, error) {
|
||||
s = strings.TrimSpace(s)
|
||||
if s == "" {
|
||||
return sql.NullInt64{}, nil
|
||||
}
|
||||
|
||||
s = strings.ToLower(s)
|
||||
|
||||
// Check for unit suffix
|
||||
multiplier := memoryUnitMultiplier(s[len(s)-1])
|
||||
if multiplier > 0 {
|
||||
numStr := s[:len(s)-1]
|
||||
|
||||
val, err := strconv.ParseFloat(numStr, 64)
|
||||
if err != nil {
|
||||
return sql.NullInt64{}, ErrInvalidMemoryFormat
|
||||
}
|
||||
|
||||
if val <= 0 {
|
||||
return sql.NullInt64{}, ErrNegativeValue
|
||||
}
|
||||
|
||||
return sql.NullInt64{Int64: int64(val * float64(multiplier)), Valid: true}, nil
|
||||
}
|
||||
|
||||
// Plain bytes
|
||||
val, err := strconv.ParseInt(s, 10, 64)
|
||||
if err != nil {
|
||||
return sql.NullInt64{}, ErrInvalidMemoryFormat
|
||||
}
|
||||
|
||||
if val <= 0 {
|
||||
return sql.NullInt64{}, ErrNegativeValue
|
||||
}
|
||||
|
||||
return sql.NullInt64{Int64: val, Valid: true}, nil
|
||||
}
|
||||
|
||||
// formatDeployKey formats an SSH public key with a descriptive comment.
|
||||
// Format: ssh-ed25519 AAAA... upaas_2025-01-15_myapp
|
||||
func formatDeployKey(pubKey string, createdAt time.Time, appName string) string {
|
||||
|
||||
@@ -1,282 +0,0 @@
|
||||
package handlers
|
||||
|
||||
import (
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"net/http"
|
||||
"time"
|
||||
|
||||
"github.com/go-chi/chi/v5"
|
||||
|
||||
"sneak.berlin/go/upaas/internal/models"
|
||||
"sneak.berlin/go/upaas/internal/service/app"
|
||||
"sneak.berlin/go/upaas/templates"
|
||||
)
|
||||
|
||||
// importMaxBodyBytes is the maximum allowed request body size for backup import (10 MB).
|
||||
const importMaxBodyBytes = 10 << 20
|
||||
|
||||
// HandleExportApp exports a single app's configuration as a JSON download.
|
||||
func (h *Handlers) HandleExportApp() http.HandlerFunc {
|
||||
return func(writer http.ResponseWriter, request *http.Request) {
|
||||
appID := chi.URLParam(request, "id")
|
||||
|
||||
application, findErr := models.FindApp(request.Context(), h.db, appID)
|
||||
if findErr != nil || application == nil {
|
||||
http.NotFound(writer, request)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
bundle, exportErr := h.appService.ExportApp(request.Context(), application)
|
||||
if exportErr != nil {
|
||||
h.log.Error("failed to export app", "error", exportErr, "app", application.Name)
|
||||
http.Error(writer, "Internal Server Error", http.StatusInternalServerError)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
filename := fmt.Sprintf("upaas-backup-%s-%s.json",
|
||||
application.Name,
|
||||
time.Now().UTC().Format("20060102-150405"),
|
||||
)
|
||||
|
||||
writer.Header().Set("Content-Type", "application/json")
|
||||
writer.Header().Set("Content-Disposition",
|
||||
`attachment; filename="`+filename+`"`)
|
||||
|
||||
encoder := json.NewEncoder(writer)
|
||||
encoder.SetIndent("", " ")
|
||||
|
||||
err := encoder.Encode(bundle)
|
||||
if err != nil {
|
||||
h.log.Error("failed to encode backup", "error", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// HandleExportAllApps exports all app configurations as a JSON download.
|
||||
func (h *Handlers) HandleExportAllApps() http.HandlerFunc {
|
||||
return func(writer http.ResponseWriter, request *http.Request) {
|
||||
bundle, exportErr := h.appService.ExportAllApps(request.Context())
|
||||
if exportErr != nil {
|
||||
h.log.Error("failed to export all apps", "error", exportErr)
|
||||
http.Error(writer, "Internal Server Error", http.StatusInternalServerError)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
filename := fmt.Sprintf("upaas-backup-all-%s.json",
|
||||
time.Now().UTC().Format("20060102-150405"),
|
||||
)
|
||||
|
||||
writer.Header().Set("Content-Type", "application/json")
|
||||
writer.Header().Set("Content-Disposition",
|
||||
`attachment; filename="`+filename+`"`)
|
||||
|
||||
encoder := json.NewEncoder(writer)
|
||||
encoder.SetIndent("", " ")
|
||||
|
||||
err := encoder.Encode(bundle)
|
||||
if err != nil {
|
||||
h.log.Error("failed to encode backup", "error", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// HandleImportPage renders the import/restore page.
|
||||
func (h *Handlers) HandleImportPage() http.HandlerFunc {
|
||||
tmpl := templates.GetParsed()
|
||||
|
||||
return func(writer http.ResponseWriter, request *http.Request) {
|
||||
data := h.addGlobals(map[string]any{
|
||||
"Success": request.URL.Query().Get("success"),
|
||||
}, request)
|
||||
|
||||
h.renderTemplate(writer, tmpl, "backup_import.html", data)
|
||||
}
|
||||
}
|
||||
|
||||
// HandleImportApps processes an uploaded backup JSON file and imports apps.
|
||||
func (h *Handlers) HandleImportApps() http.HandlerFunc {
|
||||
tmpl := templates.GetParsed()
|
||||
|
||||
return func(writer http.ResponseWriter, request *http.Request) {
|
||||
bundle, parseErr := h.parseBackupUpload(request)
|
||||
if parseErr != "" {
|
||||
data := h.addGlobals(map[string]any{"Error": parseErr}, request)
|
||||
h.renderTemplate(writer, tmpl, "backup_import.html", data)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
imported, skipped, importErr := h.appService.ImportApps(
|
||||
request.Context(), bundle,
|
||||
)
|
||||
if importErr != nil {
|
||||
h.log.Error("failed to import apps", "error", importErr)
|
||||
|
||||
data := h.addGlobals(map[string]any{
|
||||
"Error": "Import failed: " + importErr.Error(),
|
||||
}, request)
|
||||
h.renderTemplate(writer, tmpl, "backup_import.html", data)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
successMsg := fmt.Sprintf("Imported %d app(s)", len(imported))
|
||||
if len(skipped) > 0 {
|
||||
successMsg += fmt.Sprintf(", skipped %d (name conflict)", len(skipped))
|
||||
}
|
||||
|
||||
http.Redirect(writer, request,
|
||||
"/backup/import?success="+successMsg,
|
||||
http.StatusSeeOther,
|
||||
)
|
||||
}
|
||||
}
|
||||
|
||||
// parseBackupUpload extracts and validates a BackupBundle from a multipart upload.
|
||||
// Returns the bundle and an empty string on success, or nil and an error message.
|
||||
func (h *Handlers) parseBackupUpload(
|
||||
request *http.Request,
|
||||
) (*app.BackupBundle, string) {
|
||||
request.Body = http.MaxBytesReader(nil, request.Body, importMaxBodyBytes)
|
||||
|
||||
parseErr := request.ParseMultipartForm(importMaxBodyBytes)
|
||||
if parseErr != nil {
|
||||
return nil, "Failed to parse upload: " + parseErr.Error()
|
||||
}
|
||||
|
||||
file, _, openErr := request.FormFile("backup_file")
|
||||
if openErr != nil {
|
||||
return nil, "Please select a backup file to import"
|
||||
}
|
||||
|
||||
defer func() { _ = file.Close() }()
|
||||
|
||||
var bundle app.BackupBundle
|
||||
|
||||
decodeErr := json.NewDecoder(file).Decode(&bundle)
|
||||
if decodeErr != nil {
|
||||
return nil, "Invalid backup file: " + decodeErr.Error()
|
||||
}
|
||||
|
||||
if bundle.Version != 1 {
|
||||
return nil, fmt.Sprintf(
|
||||
"Unsupported backup version: %d (expected 1)", bundle.Version,
|
||||
)
|
||||
}
|
||||
|
||||
if len(bundle.Apps) == 0 {
|
||||
return nil, "Backup file contains no apps"
|
||||
}
|
||||
|
||||
return &bundle, ""
|
||||
}
|
||||
|
||||
// HandleAPIExportApp exports a single app's configuration as JSON via API.
|
||||
func (h *Handlers) HandleAPIExportApp() http.HandlerFunc {
|
||||
return func(writer http.ResponseWriter, request *http.Request) {
|
||||
appID := chi.URLParam(request, "id")
|
||||
|
||||
application, err := h.appService.GetApp(request.Context(), appID)
|
||||
if err != nil {
|
||||
h.respondJSON(writer, request,
|
||||
map[string]string{"error": "internal server error"},
|
||||
http.StatusInternalServerError)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
if application == nil {
|
||||
h.respondJSON(writer, request,
|
||||
map[string]string{"error": "app not found"},
|
||||
http.StatusNotFound)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
bundle, exportErr := h.appService.ExportApp(request.Context(), application)
|
||||
if exportErr != nil {
|
||||
h.log.Error("failed to export app", "error", exportErr)
|
||||
h.respondJSON(writer, request,
|
||||
map[string]string{"error": "failed to export app"},
|
||||
http.StatusInternalServerError)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
h.respondJSON(writer, request, bundle, http.StatusOK)
|
||||
}
|
||||
}
|
||||
|
||||
// HandleAPIExportAllApps exports all app configurations as JSON via API.
|
||||
func (h *Handlers) HandleAPIExportAllApps() http.HandlerFunc {
|
||||
return func(writer http.ResponseWriter, request *http.Request) {
|
||||
bundle, exportErr := h.appService.ExportAllApps(request.Context())
|
||||
if exportErr != nil {
|
||||
h.log.Error("failed to export all apps", "error", exportErr)
|
||||
h.respondJSON(writer, request,
|
||||
map[string]string{"error": "failed to export apps"},
|
||||
http.StatusInternalServerError)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
h.respondJSON(writer, request, bundle, http.StatusOK)
|
||||
}
|
||||
}
|
||||
|
||||
// HandleAPIImportApps imports app configurations from a JSON request body via API.
|
||||
func (h *Handlers) HandleAPIImportApps() http.HandlerFunc {
|
||||
return func(writer http.ResponseWriter, request *http.Request) {
|
||||
request.Body = http.MaxBytesReader(writer, request.Body, importMaxBodyBytes)
|
||||
|
||||
var bundle app.BackupBundle
|
||||
|
||||
decodeErr := json.NewDecoder(request.Body).Decode(&bundle)
|
||||
if decodeErr != nil {
|
||||
h.respondJSON(writer, request,
|
||||
map[string]string{"error": "invalid request body"},
|
||||
http.StatusBadRequest)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
if bundle.Version != 1 {
|
||||
h.respondJSON(writer, request,
|
||||
map[string]string{"error": fmt.Sprintf(
|
||||
"unsupported backup version: %d", bundle.Version,
|
||||
)},
|
||||
http.StatusBadRequest)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
if len(bundle.Apps) == 0 {
|
||||
h.respondJSON(writer, request,
|
||||
map[string]string{"error": "backup contains no apps"},
|
||||
http.StatusBadRequest)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
imported, skipped, importErr := h.appService.ImportApps(
|
||||
request.Context(), &bundle,
|
||||
)
|
||||
if importErr != nil {
|
||||
h.log.Error("api: failed to import apps", "error", importErr)
|
||||
h.respondJSON(writer, request,
|
||||
map[string]string{"error": "import failed: " + importErr.Error()},
|
||||
http.StatusInternalServerError)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
h.respondJSON(writer, request, map[string]any{
|
||||
"imported": imported,
|
||||
"skipped": skipped,
|
||||
}, http.StatusOK)
|
||||
}
|
||||
}
|
||||
@@ -1,582 +0,0 @@
|
||||
package handlers_test
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"context"
|
||||
"encoding/json"
|
||||
"io"
|
||||
"mime/multipart"
|
||||
"net/http"
|
||||
"net/http/httptest"
|
||||
"strings"
|
||||
"testing"
|
||||
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
|
||||
"sneak.berlin/go/upaas/internal/models"
|
||||
"sneak.berlin/go/upaas/internal/service/app"
|
||||
)
|
||||
|
||||
// createTestAppWithConfig creates an app with env vars, labels, volumes, and ports.
|
||||
func createTestAppWithConfig(
|
||||
t *testing.T,
|
||||
tc *testContext,
|
||||
name string,
|
||||
) *models.App {
|
||||
t.Helper()
|
||||
|
||||
createdApp := createTestApp(t, tc, name)
|
||||
|
||||
// Add env vars
|
||||
ev := models.NewEnvVar(tc.database)
|
||||
ev.AppID = createdApp.ID
|
||||
ev.Key = "DATABASE_URL"
|
||||
ev.Value = "postgres://localhost/mydb"
|
||||
require.NoError(t, ev.Save(context.Background()))
|
||||
|
||||
// Add label
|
||||
label := models.NewLabel(tc.database)
|
||||
label.AppID = createdApp.ID
|
||||
label.Key = "traefik.enable"
|
||||
label.Value = "true"
|
||||
require.NoError(t, label.Save(context.Background()))
|
||||
|
||||
// Add volume
|
||||
volume := models.NewVolume(tc.database)
|
||||
volume.AppID = createdApp.ID
|
||||
volume.HostPath = "/data/app"
|
||||
volume.ContainerPath = "/app/data"
|
||||
volume.ReadOnly = false
|
||||
require.NoError(t, volume.Save(context.Background()))
|
||||
|
||||
// Add port
|
||||
port := models.NewPort(tc.database)
|
||||
port.AppID = createdApp.ID
|
||||
port.HostPort = 8080
|
||||
port.ContainerPort = 80
|
||||
port.Protocol = models.PortProtocolTCP
|
||||
require.NoError(t, port.Save(context.Background()))
|
||||
|
||||
return createdApp
|
||||
}
|
||||
|
||||
// createTestAppWithConfigPort creates an app with a custom host port.
|
||||
func createTestAppWithConfigPort(
|
||||
t *testing.T,
|
||||
tc *testContext,
|
||||
name string,
|
||||
hostPort int,
|
||||
) *models.App {
|
||||
t.Helper()
|
||||
|
||||
createdApp := createTestApp(t, tc, name)
|
||||
|
||||
ev := models.NewEnvVar(tc.database)
|
||||
ev.AppID = createdApp.ID
|
||||
ev.Key = "DATABASE_URL"
|
||||
ev.Value = "postgres://localhost/mydb"
|
||||
require.NoError(t, ev.Save(context.Background()))
|
||||
|
||||
port := models.NewPort(tc.database)
|
||||
port.AppID = createdApp.ID
|
||||
port.HostPort = hostPort
|
||||
port.ContainerPort = 80
|
||||
port.Protocol = models.PortProtocolTCP
|
||||
require.NoError(t, port.Save(context.Background()))
|
||||
|
||||
return createdApp
|
||||
}
|
||||
|
||||
func TestHandleExportApp(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
createdApp := createTestAppWithConfig(t, testCtx, "export-test-app")
|
||||
|
||||
request := httptest.NewRequest(http.MethodGet, "/apps/"+createdApp.ID+"/export", nil)
|
||||
request = addChiURLParams(request, map[string]string{"id": createdApp.ID})
|
||||
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleExportApp()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
||||
assert.Contains(t, recorder.Header().Get("Content-Type"), "application/json")
|
||||
assert.Contains(t, recorder.Header().Get("Content-Disposition"), "attachment")
|
||||
assert.Contains(t, recorder.Header().Get("Content-Disposition"), "export-test-app")
|
||||
|
||||
var bundle app.BackupBundle
|
||||
require.NoError(t, json.Unmarshal(recorder.Body.Bytes(), &bundle))
|
||||
|
||||
assert.Equal(t, 1, bundle.Version)
|
||||
assert.NotEmpty(t, bundle.ExportedAt)
|
||||
require.Len(t, bundle.Apps, 1)
|
||||
|
||||
appBackup := bundle.Apps[0]
|
||||
assert.Equal(t, "export-test-app", appBackup.Name)
|
||||
assert.Equal(t, "main", appBackup.Branch)
|
||||
assert.Len(t, appBackup.EnvVars, 1)
|
||||
assert.Equal(t, "DATABASE_URL", appBackup.EnvVars[0].Key)
|
||||
assert.Equal(t, "postgres://localhost/mydb", appBackup.EnvVars[0].Value)
|
||||
assert.Len(t, appBackup.Labels, 1)
|
||||
assert.Equal(t, "traefik.enable", appBackup.Labels[0].Key)
|
||||
assert.Len(t, appBackup.Volumes, 1)
|
||||
assert.Equal(t, "/data/app", appBackup.Volumes[0].HostPath)
|
||||
assert.Len(t, appBackup.Ports, 1)
|
||||
assert.Equal(t, 8080, appBackup.Ports[0].HostPort)
|
||||
}
|
||||
|
||||
func TestHandleExportAppNotFound(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
|
||||
request := httptest.NewRequest(http.MethodGet, "/apps/nonexistent/export", nil)
|
||||
request = addChiURLParams(request, map[string]string{"id": "nonexistent"})
|
||||
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleExportApp()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
assert.Equal(t, http.StatusNotFound, recorder.Code)
|
||||
}
|
||||
|
||||
func TestHandleExportAllApps(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
createTestAppWithConfig(t, testCtx, "export-all-app1")
|
||||
createTestAppWithConfigPort(t, testCtx, "export-all-app2", 8081)
|
||||
|
||||
request := httptest.NewRequest(http.MethodGet, "/backup/export", nil)
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleExportAllApps()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
||||
assert.Contains(t, recorder.Header().Get("Content-Disposition"), "upaas-backup-all")
|
||||
|
||||
var bundle app.BackupBundle
|
||||
require.NoError(t, json.Unmarshal(recorder.Body.Bytes(), &bundle))
|
||||
|
||||
assert.Equal(t, 1, bundle.Version)
|
||||
assert.Len(t, bundle.Apps, 2)
|
||||
}
|
||||
|
||||
func TestHandleExportAllAppsEmpty(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
|
||||
request := httptest.NewRequest(http.MethodGet, "/backup/export", nil)
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleExportAllApps()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
||||
|
||||
var bundle app.BackupBundle
|
||||
require.NoError(t, json.Unmarshal(recorder.Body.Bytes(), &bundle))
|
||||
|
||||
assert.Empty(t, bundle.Apps)
|
||||
}
|
||||
|
||||
// createMultipartBackupRequest builds a multipart form request with backup JSON as a file upload.
|
||||
func createMultipartBackupRequest(
|
||||
t *testing.T,
|
||||
backupJSON string,
|
||||
) *http.Request {
|
||||
t.Helper()
|
||||
|
||||
var body bytes.Buffer
|
||||
|
||||
writer := multipart.NewWriter(&body)
|
||||
|
||||
part, err := writer.CreateFormFile("backup_file", "backup.json")
|
||||
require.NoError(t, err)
|
||||
|
||||
_, err = io.WriteString(part, backupJSON)
|
||||
require.NoError(t, err)
|
||||
|
||||
require.NoError(t, writer.Close())
|
||||
|
||||
request := httptest.NewRequest(http.MethodPost, "/backup/import", &body)
|
||||
request.Header.Set("Content-Type", writer.FormDataContentType())
|
||||
|
||||
return request
|
||||
}
|
||||
|
||||
func TestHandleImportApps(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
|
||||
backupJSON := `{
|
||||
"version": 1,
|
||||
"exportedAt": "2025-01-01T00:00:00Z",
|
||||
"apps": [{
|
||||
"name": "imported-app",
|
||||
"repoUrl": "git@example.com:user/repo.git",
|
||||
"branch": "main",
|
||||
"dockerfilePath": "Dockerfile",
|
||||
"envVars": [{"key": "FOO", "value": "bar"}],
|
||||
"labels": [{"key": "app.name", "value": "test"}],
|
||||
"volumes": [{"hostPath": "/data", "containerPath": "/app/data", "readOnly": true}],
|
||||
"ports": [{"hostPort": 3000, "containerPort": 8080, "protocol": "tcp"}]
|
||||
}]
|
||||
}`
|
||||
|
||||
request := createMultipartBackupRequest(t, backupJSON)
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleImportApps()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
// Should redirect on success
|
||||
assert.Equal(t, http.StatusSeeOther, recorder.Code)
|
||||
assert.Contains(t, recorder.Header().Get("Location"), "success=")
|
||||
|
||||
// Verify the app was created
|
||||
apps, err := models.AllApps(context.Background(), testCtx.database)
|
||||
require.NoError(t, err)
|
||||
require.Len(t, apps, 1)
|
||||
assert.Equal(t, "imported-app", apps[0].Name)
|
||||
|
||||
// Verify env vars
|
||||
envVars, _ := apps[0].GetEnvVars(context.Background())
|
||||
require.Len(t, envVars, 1)
|
||||
assert.Equal(t, "FOO", envVars[0].Key)
|
||||
assert.Equal(t, "bar", envVars[0].Value)
|
||||
|
||||
// Verify labels
|
||||
labels, _ := apps[0].GetLabels(context.Background())
|
||||
require.Len(t, labels, 1)
|
||||
assert.Equal(t, "app.name", labels[0].Key)
|
||||
|
||||
// Verify volumes
|
||||
volumes, _ := apps[0].GetVolumes(context.Background())
|
||||
require.Len(t, volumes, 1)
|
||||
assert.Equal(t, "/data", volumes[0].HostPath)
|
||||
assert.True(t, volumes[0].ReadOnly)
|
||||
|
||||
// Verify ports
|
||||
ports, _ := apps[0].GetPorts(context.Background())
|
||||
require.Len(t, ports, 1)
|
||||
assert.Equal(t, 3000, ports[0].HostPort)
|
||||
assert.Equal(t, 8080, ports[0].ContainerPort)
|
||||
}
|
||||
|
||||
func TestHandleImportAppsSkipsDuplicateNames(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
|
||||
// Create an existing app with same name
|
||||
createTestApp(t, testCtx, "existing-app")
|
||||
|
||||
backupJSON := `{
|
||||
"version": 1,
|
||||
"exportedAt": "2025-01-01T00:00:00Z",
|
||||
"apps": [
|
||||
{
|
||||
"name": "existing-app",
|
||||
"repoUrl": "git@example.com:user/repo.git",
|
||||
"branch": "main",
|
||||
"dockerfilePath": "Dockerfile",
|
||||
"envVars": [],
|
||||
"labels": [],
|
||||
"volumes": [],
|
||||
"ports": []
|
||||
},
|
||||
{
|
||||
"name": "new-app",
|
||||
"repoUrl": "git@example.com:user/new.git",
|
||||
"branch": "main",
|
||||
"dockerfilePath": "Dockerfile",
|
||||
"envVars": [],
|
||||
"labels": [],
|
||||
"volumes": [],
|
||||
"ports": []
|
||||
}
|
||||
]
|
||||
}`
|
||||
|
||||
request := createMultipartBackupRequest(t, backupJSON)
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleImportApps()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
assert.Equal(t, http.StatusSeeOther, recorder.Code)
|
||||
assert.Contains(t, recorder.Header().Get("Location"), "skipped")
|
||||
|
||||
// Should have 2 apps total (existing + new)
|
||||
apps, err := models.AllApps(context.Background(), testCtx.database)
|
||||
require.NoError(t, err)
|
||||
assert.Len(t, apps, 2)
|
||||
}
|
||||
|
||||
func TestHandleImportAppsInvalidJSON(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
|
||||
request := createMultipartBackupRequest(t, "not valid json")
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleImportApps()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
// Should render the page with error, not redirect
|
||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
||||
assert.Contains(t, recorder.Body.String(), "Invalid backup file")
|
||||
}
|
||||
|
||||
func TestHandleImportAppsUnsupportedVersion(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
|
||||
backupJSON := `{"version": 99, "exportedAt": "2025-01-01T00:00:00Z", "apps": [{"name": "test"}]}`
|
||||
|
||||
request := createMultipartBackupRequest(t, backupJSON)
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleImportApps()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
||||
assert.Contains(t, recorder.Body.String(), "Unsupported backup version")
|
||||
}
|
||||
|
||||
func TestHandleImportAppsEmptyBundle(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
|
||||
backupJSON := `{"version": 1, "exportedAt": "2025-01-01T00:00:00Z", "apps": []}`
|
||||
|
||||
request := createMultipartBackupRequest(t, backupJSON)
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleImportApps()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
||||
assert.Contains(t, recorder.Body.String(), "contains no apps")
|
||||
}
|
||||
|
||||
func TestHandleImportPage(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
|
||||
request := httptest.NewRequest(http.MethodGet, "/backup/import", nil)
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
handler := testCtx.handlers.HandleImportPage()
|
||||
handler.ServeHTTP(recorder, request)
|
||||
|
||||
assert.Equal(t, http.StatusOK, recorder.Code)
|
||||
assert.Contains(t, recorder.Body.String(), "Import Backup")
|
||||
}
|
||||
|
||||
func TestExportImportRoundTrip(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testCtx := setupTestHandlers(t)
|
||||
createTestAppWithConfig(t, testCtx, "roundtrip-app")
|
||||
|
||||
// Export
|
||||
exportReq := httptest.NewRequest(http.MethodGet, "/backup/export", nil)
|
||||
exportRec := httptest.NewRecorder()
|
||||
|
||||
testCtx.handlers.HandleExportAllApps().ServeHTTP(exportRec, exportReq)
|
||||
|
||||
require.Equal(t, http.StatusOK, exportRec.Code)
|
||||
|
||||
exportedJSON := exportRec.Body.String()
|
||||
|
||||
// Delete the original app
|
||||
apps, _ := models.AllApps(context.Background(), testCtx.database)
|
||||
for _, a := range apps {
|
||||
require.NoError(t, a.Delete(context.Background()))
|
||||
}
|
||||
|
||||
// Import
|
||||
importReq := createMultipartBackupRequest(t, exportedJSON)
|
||||
importRec := httptest.NewRecorder()
|
||||
|
||||
testCtx.handlers.HandleImportApps().ServeHTTP(importRec, importReq)
|
||||
|
||||
assert.Equal(t, http.StatusSeeOther, importRec.Code)
|
||||
|
||||
// Verify the app was recreated with all config
|
||||
restoredApps, _ := models.AllApps(context.Background(), testCtx.database)
|
||||
require.Len(t, restoredApps, 1)
|
||||
assert.Equal(t, "roundtrip-app", restoredApps[0].Name)
|
||||
|
||||
envVars, _ := restoredApps[0].GetEnvVars(context.Background())
|
||||
assert.Len(t, envVars, 1)
|
||||
|
||||
labels, _ := restoredApps[0].GetLabels(context.Background())
|
||||
assert.Len(t, labels, 1)
|
||||
|
||||
volumes, _ := restoredApps[0].GetVolumes(context.Background())
|
||||
assert.Len(t, volumes, 1)
|
||||
|
||||
ports, _ := restoredApps[0].GetPorts(context.Background())
|
||||
assert.Len(t, ports, 1)
|
||||
}
|
||||
|
||||
// TestAPIExportApp tests the API endpoint for exporting a single app.
|
||||
func TestAPIExportApp(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
tc, cookies := setupAPITest(t)
|
||||
|
||||
createdApp, err := tc.appSvc.CreateApp(t.Context(), app.CreateAppInput{
|
||||
Name: "api-export-app",
|
||||
RepoURL: "git@example.com:user/repo.git",
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
rr := apiGet(t, tc, cookies, "/api/v1/apps/"+createdApp.ID+"/export")
|
||||
assert.Equal(t, http.StatusOK, rr.Code)
|
||||
|
||||
var bundle app.BackupBundle
|
||||
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &bundle))
|
||||
|
||||
assert.Equal(t, 1, bundle.Version)
|
||||
require.Len(t, bundle.Apps, 1)
|
||||
assert.Equal(t, "api-export-app", bundle.Apps[0].Name)
|
||||
}
|
||||
|
||||
// TestAPIExportAppNotFound tests the API endpoint for a nonexistent app.
|
||||
func TestAPIExportAppNotFound(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
tc, cookies := setupAPITest(t)
|
||||
|
||||
rr := apiGet(t, tc, cookies, "/api/v1/apps/nonexistent/export")
|
||||
assert.Equal(t, http.StatusNotFound, rr.Code)
|
||||
}
|
||||
|
||||
// TestAPIExportAllApps tests the API endpoint for exporting all apps.
|
||||
func TestAPIExportAllApps(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
tc, cookies := setupAPITest(t)
|
||||
|
||||
_, err := tc.appSvc.CreateApp(t.Context(), app.CreateAppInput{
|
||||
Name: "api-export-all-1",
|
||||
RepoURL: "git@example.com:user/repo1.git",
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
_, err = tc.appSvc.CreateApp(t.Context(), app.CreateAppInput{
|
||||
Name: "api-export-all-2",
|
||||
RepoURL: "git@example.com:user/repo2.git",
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
rr := apiGet(t, tc, cookies, "/api/v1/backup/export")
|
||||
assert.Equal(t, http.StatusOK, rr.Code)
|
||||
|
||||
var bundle app.BackupBundle
|
||||
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &bundle))
|
||||
|
||||
assert.Len(t, bundle.Apps, 2)
|
||||
}
|
||||
|
||||
// TestAPIImportApps tests the API import endpoint.
|
||||
func TestAPIImportApps(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
tc, cookies := setupAPITest(t)
|
||||
|
||||
backupJSON := `{
|
||||
"version": 1,
|
||||
"exportedAt": "2025-01-01T00:00:00Z",
|
||||
"apps": [{
|
||||
"name": "api-imported-app",
|
||||
"repoUrl": "git@example.com:user/repo.git",
|
||||
"branch": "main",
|
||||
"dockerfilePath": "Dockerfile",
|
||||
"envVars": [],
|
||||
"labels": [],
|
||||
"volumes": [],
|
||||
"ports": []
|
||||
}]
|
||||
}`
|
||||
|
||||
r := apiRouter(tc)
|
||||
|
||||
req := httptest.NewRequest(http.MethodPost, "/api/v1/backup/import", strings.NewReader(backupJSON))
|
||||
req.Header.Set("Content-Type", "application/json")
|
||||
|
||||
for _, c := range cookies {
|
||||
req.AddCookie(c)
|
||||
}
|
||||
|
||||
rr := httptest.NewRecorder()
|
||||
r.ServeHTTP(rr, req)
|
||||
|
||||
assert.Equal(t, http.StatusOK, rr.Code)
|
||||
|
||||
var resp map[string]any
|
||||
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &resp))
|
||||
|
||||
imported, ok := resp["imported"].([]any)
|
||||
require.True(t, ok)
|
||||
assert.Len(t, imported, 1)
|
||||
assert.Equal(t, "api-imported-app", imported[0])
|
||||
}
|
||||
|
||||
// TestAPIImportAppsInvalidBody tests that the API rejects bad JSON.
|
||||
func TestAPIImportAppsInvalidBody(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
tc, cookies := setupAPITest(t)
|
||||
|
||||
r := apiRouter(tc)
|
||||
|
||||
req := httptest.NewRequest(http.MethodPost, "/api/v1/backup/import", strings.NewReader("not json"))
|
||||
req.Header.Set("Content-Type", "application/json")
|
||||
|
||||
for _, c := range cookies {
|
||||
req.AddCookie(c)
|
||||
}
|
||||
|
||||
rr := httptest.NewRecorder()
|
||||
r.ServeHTTP(rr, req)
|
||||
|
||||
assert.Equal(t, http.StatusBadRequest, rr.Code)
|
||||
}
|
||||
|
||||
// TestAPIImportAppsUnsupportedVersion tests that the API rejects bad versions.
|
||||
func TestAPIImportAppsUnsupportedVersion(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
tc, cookies := setupAPITest(t)
|
||||
|
||||
r := apiRouter(tc)
|
||||
|
||||
body := `{"version": 42, "apps": [{"name": "x"}]}`
|
||||
req := httptest.NewRequest(http.MethodPost, "/api/v1/backup/import", strings.NewReader(body))
|
||||
req.Header.Set("Content-Type", "application/json")
|
||||
|
||||
for _, c := range cookies {
|
||||
req.AddCookie(c)
|
||||
}
|
||||
|
||||
rr := httptest.NewRecorder()
|
||||
r.ServeHTTP(rr, req)
|
||||
|
||||
assert.Equal(t, http.StatusBadRequest, rr.Code)
|
||||
}
|
||||
195
internal/handlers/resource_limits_test.go
Normal file
195
internal/handlers/resource_limits_test.go
Normal file
@@ -0,0 +1,195 @@
|
||||
package handlers //nolint:testpackage // tests unexported parsing functions
|
||||
|
||||
import (
|
||||
"database/sql"
|
||||
"testing"
|
||||
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
func TestParseOptionalFloat64(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
t.Run("empty string returns invalid", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalFloat64("")
|
||||
require.NoError(t, err)
|
||||
assert.False(t, result.Valid)
|
||||
})
|
||||
|
||||
t.Run("whitespace only returns invalid", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalFloat64(" ")
|
||||
require.NoError(t, err)
|
||||
assert.False(t, result.Valid)
|
||||
})
|
||||
|
||||
t.Run("valid float", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalFloat64("0.5")
|
||||
require.NoError(t, err)
|
||||
assert.True(t, result.Valid)
|
||||
assert.InDelta(t, 0.5, result.Float64, 0.001)
|
||||
})
|
||||
|
||||
t.Run("valid integer", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalFloat64("2")
|
||||
require.NoError(t, err)
|
||||
assert.True(t, result.Valid)
|
||||
assert.InDelta(t, 2.0, result.Float64, 0.001)
|
||||
})
|
||||
|
||||
t.Run("negative value rejected", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
_, err := parseOptionalFloat64("-1")
|
||||
require.Error(t, err)
|
||||
})
|
||||
|
||||
t.Run("zero value rejected", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
_, err := parseOptionalFloat64("0")
|
||||
require.Error(t, err)
|
||||
})
|
||||
|
||||
t.Run("non-numeric rejected", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
_, err := parseOptionalFloat64("abc")
|
||||
require.Error(t, err)
|
||||
})
|
||||
}
|
||||
|
||||
func TestParseOptionalMemoryBytes(t *testing.T) { //nolint:funlen // table-driven test
|
||||
t.Parallel()
|
||||
|
||||
t.Run("empty string returns invalid", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes("")
|
||||
require.NoError(t, err)
|
||||
assert.False(t, result.Valid)
|
||||
})
|
||||
|
||||
t.Run("whitespace only returns invalid", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes(" ")
|
||||
require.NoError(t, err)
|
||||
assert.False(t, result.Valid)
|
||||
})
|
||||
|
||||
t.Run("plain bytes", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes("536870912")
|
||||
require.NoError(t, err)
|
||||
assert.True(t, result.Valid)
|
||||
assert.Equal(t, int64(536870912), result.Int64)
|
||||
})
|
||||
|
||||
t.Run("megabytes suffix", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes("256m")
|
||||
require.NoError(t, err)
|
||||
assert.True(t, result.Valid)
|
||||
assert.Equal(t, int64(256*1024*1024), result.Int64)
|
||||
})
|
||||
|
||||
t.Run("megabytes suffix uppercase", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes("256M")
|
||||
require.NoError(t, err)
|
||||
assert.True(t, result.Valid)
|
||||
assert.Equal(t, int64(256*1024*1024), result.Int64)
|
||||
})
|
||||
|
||||
t.Run("gigabytes suffix", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes("1g")
|
||||
require.NoError(t, err)
|
||||
assert.True(t, result.Valid)
|
||||
assert.Equal(t, int64(1024*1024*1024), result.Int64)
|
||||
})
|
||||
|
||||
t.Run("kilobytes suffix", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes("512k")
|
||||
require.NoError(t, err)
|
||||
assert.True(t, result.Valid)
|
||||
assert.Equal(t, int64(512*1024), result.Int64)
|
||||
})
|
||||
|
||||
t.Run("fractional gigabytes", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes("1.5g")
|
||||
require.NoError(t, err)
|
||||
assert.True(t, result.Valid)
|
||||
assert.Equal(t, int64(1.5*1024*1024*1024), result.Int64)
|
||||
})
|
||||
|
||||
t.Run("negative value rejected", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
_, err := parseOptionalMemoryBytes("-256m")
|
||||
require.Error(t, err)
|
||||
})
|
||||
|
||||
t.Run("zero value rejected", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
_, err := parseOptionalMemoryBytes("0")
|
||||
require.Error(t, err)
|
||||
})
|
||||
|
||||
t.Run("invalid string rejected", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
_, err := parseOptionalMemoryBytes("abc")
|
||||
require.Error(t, err)
|
||||
})
|
||||
|
||||
t.Run("negative plain bytes rejected", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
_, err := parseOptionalMemoryBytes("-100")
|
||||
require.Error(t, err)
|
||||
})
|
||||
}
|
||||
|
||||
func TestAppResourceLimitsRoundTrip(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
// Test that parsing and formatting are consistent
|
||||
tests := []struct {
|
||||
input string
|
||||
expected sql.NullInt64
|
||||
format string
|
||||
}{
|
||||
{"256m", sql.NullInt64{Int64: 256 * 1024 * 1024, Valid: true}, "256m"},
|
||||
{"1g", sql.NullInt64{Int64: 1024 * 1024 * 1024, Valid: true}, "1g"},
|
||||
{"512k", sql.NullInt64{Int64: 512 * 1024, Valid: true}, "512k"},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.input, func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
result, err := parseOptionalMemoryBytes(tt.input)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, tt.expected, result)
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -14,7 +14,8 @@ import (
|
||||
const appColumns = `id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
||||
ssh_private_key, ssh_public_key, image_id, status,
|
||||
docker_network, ntfy_topic, slack_webhook, webhook_secret_hash,
|
||||
previous_image_id, created_at, updated_at`
|
||||
previous_image_id, cpu_limit, memory_limit,
|
||||
created_at, updated_at`
|
||||
|
||||
// AppStatus represents the status of an app.
|
||||
type AppStatus string
|
||||
@@ -47,6 +48,8 @@ type App struct {
|
||||
DockerNetwork sql.NullString
|
||||
NtfyTopic sql.NullString
|
||||
SlackWebhook sql.NullString
|
||||
CPULimit sql.NullFloat64
|
||||
MemoryLimit sql.NullInt64
|
||||
CreatedAt time.Time
|
||||
UpdatedAt time.Time
|
||||
}
|
||||
@@ -142,14 +145,14 @@ func (a *App) insert(ctx context.Context) error {
|
||||
id, name, repo_url, branch, dockerfile_path, webhook_secret,
|
||||
ssh_private_key, ssh_public_key, image_id, status,
|
||||
docker_network, ntfy_topic, slack_webhook, webhook_secret_hash,
|
||||
previous_image_id
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)`
|
||||
previous_image_id, cpu_limit, memory_limit
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)`
|
||||
|
||||
_, err := a.db.Exec(ctx, query,
|
||||
a.ID, a.Name, a.RepoURL, a.Branch, a.DockerfilePath, a.WebhookSecret,
|
||||
a.SSHPrivateKey, a.SSHPublicKey, a.ImageID, a.Status,
|
||||
a.DockerNetwork, a.NtfyTopic, a.SlackWebhook, a.WebhookSecretHash,
|
||||
a.PreviousImageID,
|
||||
a.PreviousImageID, a.CPULimit, a.MemoryLimit,
|
||||
)
|
||||
if err != nil {
|
||||
return err
|
||||
@@ -165,6 +168,7 @@ func (a *App) update(ctx context.Context) error {
|
||||
image_id = ?, status = ?,
|
||||
docker_network = ?, ntfy_topic = ?, slack_webhook = ?,
|
||||
previous_image_id = ?,
|
||||
cpu_limit = ?, memory_limit = ?,
|
||||
updated_at = CURRENT_TIMESTAMP
|
||||
WHERE id = ?`
|
||||
|
||||
@@ -173,6 +177,7 @@ func (a *App) update(ctx context.Context) error {
|
||||
a.ImageID, a.Status,
|
||||
a.DockerNetwork, a.NtfyTopic, a.SlackWebhook,
|
||||
a.PreviousImageID,
|
||||
a.CPULimit, a.MemoryLimit,
|
||||
a.ID,
|
||||
)
|
||||
|
||||
@@ -188,6 +193,7 @@ func (a *App) scan(row *sql.Row) error {
|
||||
&a.DockerNetwork, &a.NtfyTopic, &a.SlackWebhook,
|
||||
&a.WebhookSecretHash,
|
||||
&a.PreviousImageID,
|
||||
&a.CPULimit, &a.MemoryLimit,
|
||||
&a.CreatedAt, &a.UpdatedAt,
|
||||
)
|
||||
}
|
||||
@@ -206,6 +212,7 @@ func scanApps(appDB *database.Database, rows *sql.Rows) ([]*App, error) {
|
||||
&app.DockerNetwork, &app.NtfyTopic, &app.SlackWebhook,
|
||||
&app.WebhookSecretHash,
|
||||
&app.PreviousImageID,
|
||||
&app.CPULimit, &app.MemoryLimit,
|
||||
&app.CreatedAt, &app.UpdatedAt,
|
||||
)
|
||||
if scanErr != nil {
|
||||
|
||||
@@ -781,6 +781,96 @@ func TestCascadeDelete(t *testing.T) {
|
||||
})
|
||||
}
|
||||
|
||||
// Resource Limits Tests.
|
||||
|
||||
func TestAppResourceLimits(t *testing.T) { //nolint:funlen // integration test with multiple subtests
|
||||
t.Parallel()
|
||||
|
||||
t.Run("saves and loads CPU limit", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testDB, cleanup := setupTestDB(t)
|
||||
defer cleanup()
|
||||
|
||||
app := createTestApp(t, testDB)
|
||||
|
||||
app.CPULimit = sql.NullFloat64{Float64: 0.5, Valid: true}
|
||||
|
||||
err := app.Save(context.Background())
|
||||
require.NoError(t, err)
|
||||
|
||||
found, err := models.FindApp(context.Background(), testDB, app.ID)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, found)
|
||||
assert.True(t, found.CPULimit.Valid)
|
||||
assert.InDelta(t, 0.5, found.CPULimit.Float64, 0.001)
|
||||
})
|
||||
|
||||
t.Run("saves and loads memory limit", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testDB, cleanup := setupTestDB(t)
|
||||
defer cleanup()
|
||||
|
||||
app := createTestApp(t, testDB)
|
||||
|
||||
app.MemoryLimit = sql.NullInt64{Int64: 536870912, Valid: true} // 512m
|
||||
|
||||
err := app.Save(context.Background())
|
||||
require.NoError(t, err)
|
||||
|
||||
found, err := models.FindApp(context.Background(), testDB, app.ID)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, found)
|
||||
assert.True(t, found.MemoryLimit.Valid)
|
||||
assert.Equal(t, int64(536870912), found.MemoryLimit.Int64)
|
||||
})
|
||||
|
||||
t.Run("null limits by default", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testDB, cleanup := setupTestDB(t)
|
||||
defer cleanup()
|
||||
|
||||
app := createTestApp(t, testDB)
|
||||
|
||||
found, err := models.FindApp(context.Background(), testDB, app.ID)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, found)
|
||||
assert.False(t, found.CPULimit.Valid)
|
||||
assert.False(t, found.MemoryLimit.Valid)
|
||||
})
|
||||
|
||||
t.Run("clears limits when set to null", func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
testDB, cleanup := setupTestDB(t)
|
||||
defer cleanup()
|
||||
|
||||
app := createTestApp(t, testDB)
|
||||
|
||||
// Set limits
|
||||
app.CPULimit = sql.NullFloat64{Float64: 1.0, Valid: true}
|
||||
app.MemoryLimit = sql.NullInt64{Int64: 1073741824, Valid: true} // 1g
|
||||
|
||||
err := app.Save(context.Background())
|
||||
require.NoError(t, err)
|
||||
|
||||
// Clear limits
|
||||
app.CPULimit = sql.NullFloat64{}
|
||||
app.MemoryLimit = sql.NullInt64{}
|
||||
|
||||
err = app.Save(context.Background())
|
||||
require.NoError(t, err)
|
||||
|
||||
found, err := models.FindApp(context.Background(), testDB, app.ID)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, found)
|
||||
assert.False(t, found.CPULimit.Valid)
|
||||
assert.False(t, found.MemoryLimit.Valid)
|
||||
})
|
||||
}
|
||||
|
||||
// Helper function to create a test app.
|
||||
func createTestApp(t *testing.T, testDB *database.Database) *models.App {
|
||||
t.Helper()
|
||||
|
||||
@@ -98,12 +98,6 @@ func (s *Server) SetupRoutes() {
|
||||
// Ports
|
||||
r.Post("/apps/{id}/ports", s.handlers.HandlePortAdd())
|
||||
r.Post("/apps/{id}/ports/{portID}/delete", s.handlers.HandlePortDelete())
|
||||
|
||||
// Backup/Restore
|
||||
r.Get("/apps/{id}/export", s.handlers.HandleExportApp())
|
||||
r.Get("/backup/export", s.handlers.HandleExportAllApps())
|
||||
r.Get("/backup/import", s.handlers.HandleImportPage())
|
||||
r.Post("/backup/import", s.handlers.HandleImportApps())
|
||||
})
|
||||
})
|
||||
|
||||
@@ -121,11 +115,6 @@ func (s *Server) SetupRoutes() {
|
||||
r.Get("/apps", s.handlers.HandleAPIListApps())
|
||||
r.Get("/apps/{id}", s.handlers.HandleAPIGetApp())
|
||||
r.Get("/apps/{id}/deployments", s.handlers.HandleAPIListDeployments())
|
||||
|
||||
// Backup/Restore API
|
||||
r.Get("/apps/{id}/export", s.handlers.HandleAPIExportApp())
|
||||
r.Get("/backup/export", s.handlers.HandleAPIExportAllApps())
|
||||
r.Post("/backup/import", s.handlers.HandleAPIImportApps())
|
||||
})
|
||||
})
|
||||
|
||||
|
||||
@@ -1,391 +0,0 @@
|
||||
package app
|
||||
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
"time"
|
||||
|
||||
"sneak.berlin/go/upaas/internal/models"
|
||||
)
|
||||
|
||||
// BackupEnvVar represents an environment variable in a backup.
|
||||
type BackupEnvVar struct {
|
||||
Key string `json:"key"`
|
||||
Value string `json:"value"`
|
||||
}
|
||||
|
||||
// BackupLabel represents a Docker label in a backup.
|
||||
type BackupLabel struct {
|
||||
Key string `json:"key"`
|
||||
Value string `json:"value"`
|
||||
}
|
||||
|
||||
// BackupVolume represents a volume mount in a backup.
|
||||
type BackupVolume struct {
|
||||
HostPath string `json:"hostPath"`
|
||||
ContainerPath string `json:"containerPath"`
|
||||
ReadOnly bool `json:"readOnly"`
|
||||
}
|
||||
|
||||
// BackupPort represents a port mapping in a backup.
|
||||
type BackupPort struct {
|
||||
HostPort int `json:"hostPort"`
|
||||
ContainerPort int `json:"containerPort"`
|
||||
Protocol string `json:"protocol"`
|
||||
}
|
||||
|
||||
// Backup represents the exported configuration of a single app.
|
||||
type Backup struct {
|
||||
Name string `json:"name"`
|
||||
RepoURL string `json:"repoUrl"`
|
||||
Branch string `json:"branch"`
|
||||
DockerfilePath string `json:"dockerfilePath"`
|
||||
DockerNetwork string `json:"dockerNetwork,omitempty"`
|
||||
NtfyTopic string `json:"ntfyTopic,omitempty"`
|
||||
SlackWebhook string `json:"slackWebhook,omitempty"`
|
||||
EnvVars []BackupEnvVar `json:"envVars"`
|
||||
Labels []BackupLabel `json:"labels"`
|
||||
Volumes []BackupVolume `json:"volumes"`
|
||||
Ports []BackupPort `json:"ports"`
|
||||
}
|
||||
|
||||
// BackupBundle represents a complete backup of one or more apps.
|
||||
type BackupBundle struct {
|
||||
Version int `json:"version"`
|
||||
ExportedAt string `json:"exportedAt"`
|
||||
Apps []Backup `json:"apps"`
|
||||
}
|
||||
|
||||
// backupVersion is the current backup format version.
|
||||
const backupVersion = 1
|
||||
|
||||
// ExportApp exports a single app's configuration as a BackupBundle.
|
||||
func (svc *Service) ExportApp(
|
||||
ctx context.Context,
|
||||
application *models.App,
|
||||
) (*BackupBundle, error) {
|
||||
appBackup, err := svc.buildAppBackup(ctx, application)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
return &BackupBundle{
|
||||
Version: backupVersion,
|
||||
ExportedAt: time.Now().UTC().Format(time.RFC3339),
|
||||
Apps: []Backup{appBackup},
|
||||
}, nil
|
||||
}
|
||||
|
||||
// ExportAllApps exports all app configurations as a BackupBundle.
|
||||
func (svc *Service) ExportAllApps(ctx context.Context) (*BackupBundle, error) {
|
||||
apps, err := models.AllApps(ctx, svc.db)
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("listing apps for export: %w", err)
|
||||
}
|
||||
|
||||
backups := make([]Backup, 0, len(apps))
|
||||
|
||||
for _, application := range apps {
|
||||
appBackup, buildErr := svc.buildAppBackup(ctx, application)
|
||||
if buildErr != nil {
|
||||
return nil, buildErr
|
||||
}
|
||||
|
||||
backups = append(backups, appBackup)
|
||||
}
|
||||
|
||||
return &BackupBundle{
|
||||
Version: backupVersion,
|
||||
ExportedAt: time.Now().UTC().Format(time.RFC3339),
|
||||
Apps: backups,
|
||||
}, nil
|
||||
}
|
||||
|
||||
// ImportApps imports app configurations from a BackupBundle.
|
||||
// It creates new apps (with fresh IDs, SSH keys, and webhook secrets)
|
||||
// and populates their env vars, labels, volumes, and ports.
|
||||
// Apps whose names conflict with existing apps are skipped and reported.
|
||||
func (svc *Service) ImportApps(
|
||||
ctx context.Context,
|
||||
bundle *BackupBundle,
|
||||
) ([]string, []string, error) {
|
||||
// Build a set of existing app names for conflict detection
|
||||
existingApps, listErr := models.AllApps(ctx, svc.db)
|
||||
if listErr != nil {
|
||||
return nil, nil, fmt.Errorf("listing existing apps: %w", listErr)
|
||||
}
|
||||
|
||||
existingNames := make(map[string]bool, len(existingApps))
|
||||
for _, a := range existingApps {
|
||||
existingNames[a.Name] = true
|
||||
}
|
||||
|
||||
var imported, skipped []string
|
||||
|
||||
for _, ab := range bundle.Apps {
|
||||
if existingNames[ab.Name] {
|
||||
skipped = append(skipped, ab.Name)
|
||||
|
||||
continue
|
||||
}
|
||||
|
||||
importErr := svc.importSingleApp(ctx, ab)
|
||||
if importErr != nil {
|
||||
return imported, skipped, fmt.Errorf(
|
||||
"importing app %q: %w", ab.Name, importErr,
|
||||
)
|
||||
}
|
||||
|
||||
imported = append(imported, ab.Name)
|
||||
}
|
||||
|
||||
return imported, skipped, nil
|
||||
}
|
||||
|
||||
// importSingleApp creates a single app from backup data.
|
||||
func (svc *Service) importSingleApp(
|
||||
ctx context.Context,
|
||||
ab Backup,
|
||||
) error {
|
||||
createdApp, createErr := svc.CreateApp(ctx, CreateAppInput{
|
||||
Name: ab.Name,
|
||||
RepoURL: ab.RepoURL,
|
||||
Branch: ab.Branch,
|
||||
DockerfilePath: ab.DockerfilePath,
|
||||
DockerNetwork: ab.DockerNetwork,
|
||||
NtfyTopic: ab.NtfyTopic,
|
||||
SlackWebhook: ab.SlackWebhook,
|
||||
})
|
||||
if createErr != nil {
|
||||
return fmt.Errorf("creating app: %w", createErr)
|
||||
}
|
||||
|
||||
envErr := svc.importEnvVars(ctx, createdApp.ID, ab.EnvVars)
|
||||
if envErr != nil {
|
||||
return envErr
|
||||
}
|
||||
|
||||
labelErr := svc.importLabels(ctx, createdApp.ID, ab.Labels)
|
||||
if labelErr != nil {
|
||||
return labelErr
|
||||
}
|
||||
|
||||
volErr := svc.importVolumes(ctx, createdApp.ID, ab.Volumes)
|
||||
if volErr != nil {
|
||||
return volErr
|
||||
}
|
||||
|
||||
portErr := svc.importPorts(ctx, createdApp.ID, ab.Ports)
|
||||
if portErr != nil {
|
||||
return portErr
|
||||
}
|
||||
|
||||
svc.log.Info("app imported from backup",
|
||||
"id", createdApp.ID, "name", createdApp.Name)
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// importEnvVars adds env vars from backup to an app.
|
||||
func (svc *Service) importEnvVars(
|
||||
ctx context.Context,
|
||||
appID string,
|
||||
envVars []BackupEnvVar,
|
||||
) error {
|
||||
for _, ev := range envVars {
|
||||
addErr := svc.AddEnvVar(ctx, appID, ev.Key, ev.Value)
|
||||
if addErr != nil {
|
||||
return fmt.Errorf("adding env var %q: %w", ev.Key, addErr)
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// importLabels adds labels from backup to an app.
|
||||
func (svc *Service) importLabels(
|
||||
ctx context.Context,
|
||||
appID string,
|
||||
labels []BackupLabel,
|
||||
) error {
|
||||
for _, l := range labels {
|
||||
addErr := svc.AddLabel(ctx, appID, l.Key, l.Value)
|
||||
if addErr != nil {
|
||||
return fmt.Errorf("adding label %q: %w", l.Key, addErr)
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// importVolumes adds volumes from backup to an app.
|
||||
func (svc *Service) importVolumes(
|
||||
ctx context.Context,
|
||||
appID string,
|
||||
volumes []BackupVolume,
|
||||
) error {
|
||||
for _, v := range volumes {
|
||||
addErr := svc.AddVolume(ctx, appID, v.HostPath, v.ContainerPath, v.ReadOnly)
|
||||
if addErr != nil {
|
||||
return fmt.Errorf("adding volume %q: %w", v.ContainerPath, addErr)
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// importPorts adds ports from backup to an app.
|
||||
func (svc *Service) importPorts(
|
||||
ctx context.Context,
|
||||
appID string,
|
||||
ports []BackupPort,
|
||||
) error {
|
||||
for _, p := range ports {
|
||||
port := models.NewPort(svc.db)
|
||||
port.AppID = appID
|
||||
port.HostPort = p.HostPort
|
||||
port.ContainerPort = p.ContainerPort
|
||||
port.Protocol = models.PortProtocol(p.Protocol)
|
||||
|
||||
if port.Protocol == "" {
|
||||
port.Protocol = models.PortProtocolTCP
|
||||
}
|
||||
|
||||
saveErr := port.Save(ctx)
|
||||
if saveErr != nil {
|
||||
return fmt.Errorf("adding port %d: %w", p.HostPort, saveErr)
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// buildAppBackup collects all configuration for a single app into a Backup.
|
||||
func (svc *Service) buildAppBackup(
|
||||
ctx context.Context,
|
||||
application *models.App,
|
||||
) (Backup, error) {
|
||||
envVars, labels, volumes, ports, err := svc.fetchAppResources(ctx, application)
|
||||
if err != nil {
|
||||
return Backup{}, err
|
||||
}
|
||||
|
||||
backup := Backup{
|
||||
Name: application.Name,
|
||||
RepoURL: application.RepoURL,
|
||||
Branch: application.Branch,
|
||||
DockerfilePath: application.DockerfilePath,
|
||||
EnvVars: convertEnvVars(envVars),
|
||||
Labels: convertLabels(labels),
|
||||
Volumes: convertVolumes(volumes),
|
||||
Ports: convertPorts(ports),
|
||||
}
|
||||
|
||||
if application.DockerNetwork.Valid {
|
||||
backup.DockerNetwork = application.DockerNetwork.String
|
||||
}
|
||||
|
||||
if application.NtfyTopic.Valid {
|
||||
backup.NtfyTopic = application.NtfyTopic.String
|
||||
}
|
||||
|
||||
if application.SlackWebhook.Valid {
|
||||
backup.SlackWebhook = application.SlackWebhook.String
|
||||
}
|
||||
|
||||
return backup, nil
|
||||
}
|
||||
|
||||
// fetchAppResources retrieves all sub-resources for an app.
|
||||
func (svc *Service) fetchAppResources(
|
||||
ctx context.Context,
|
||||
application *models.App,
|
||||
) ([]*models.EnvVar, []*models.Label, []*models.Volume, []*models.Port, error) {
|
||||
envVars, err := application.GetEnvVars(ctx)
|
||||
if err != nil {
|
||||
return nil, nil, nil, nil, fmt.Errorf(
|
||||
"getting env vars for %q: %w", application.Name, err,
|
||||
)
|
||||
}
|
||||
|
||||
labels, err := application.GetLabels(ctx)
|
||||
if err != nil {
|
||||
return nil, nil, nil, nil, fmt.Errorf(
|
||||
"getting labels for %q: %w", application.Name, err,
|
||||
)
|
||||
}
|
||||
|
||||
volumes, err := application.GetVolumes(ctx)
|
||||
if err != nil {
|
||||
return nil, nil, nil, nil, fmt.Errorf(
|
||||
"getting volumes for %q: %w", application.Name, err,
|
||||
)
|
||||
}
|
||||
|
||||
ports, err := application.GetPorts(ctx)
|
||||
if err != nil {
|
||||
return nil, nil, nil, nil, fmt.Errorf(
|
||||
"getting ports for %q: %w", application.Name, err,
|
||||
)
|
||||
}
|
||||
|
||||
return envVars, labels, volumes, ports, nil
|
||||
}
|
||||
|
||||
// convertEnvVars converts model env vars to backup format.
|
||||
func convertEnvVars(envVars []*models.EnvVar) []BackupEnvVar {
|
||||
result := make([]BackupEnvVar, 0, len(envVars))
|
||||
|
||||
for _, ev := range envVars {
|
||||
result = append(result, BackupEnvVar{
|
||||
Key: ev.Key,
|
||||
Value: ev.Value,
|
||||
})
|
||||
}
|
||||
|
||||
return result
|
||||
}
|
||||
|
||||
// convertLabels converts model labels to backup format.
|
||||
func convertLabels(labels []*models.Label) []BackupLabel {
|
||||
result := make([]BackupLabel, 0, len(labels))
|
||||
|
||||
for _, l := range labels {
|
||||
result = append(result, BackupLabel{
|
||||
Key: l.Key,
|
||||
Value: l.Value,
|
||||
})
|
||||
}
|
||||
|
||||
return result
|
||||
}
|
||||
|
||||
// convertVolumes converts model volumes to backup format.
|
||||
func convertVolumes(volumes []*models.Volume) []BackupVolume {
|
||||
result := make([]BackupVolume, 0, len(volumes))
|
||||
|
||||
for _, v := range volumes {
|
||||
result = append(result, BackupVolume{
|
||||
HostPath: v.HostPath,
|
||||
ContainerPath: v.ContainerPath,
|
||||
ReadOnly: v.ReadOnly,
|
||||
})
|
||||
}
|
||||
|
||||
return result
|
||||
}
|
||||
|
||||
// convertPorts converts model ports to backup format.
|
||||
func convertPorts(ports []*models.Port) []BackupPort {
|
||||
result := make([]BackupPort, 0, len(ports))
|
||||
|
||||
for _, p := range ports {
|
||||
result = append(result, BackupPort{
|
||||
HostPort: p.HostPort,
|
||||
ContainerPort: p.ContainerPort,
|
||||
Protocol: string(p.Protocol),
|
||||
})
|
||||
}
|
||||
|
||||
return result
|
||||
}
|
||||
@@ -1,379 +0,0 @@
|
||||
package app_test
|
||||
|
||||
import (
|
||||
"context"
|
||||
"testing"
|
||||
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
"go.uber.org/fx"
|
||||
|
||||
"sneak.berlin/go/upaas/internal/config"
|
||||
"sneak.berlin/go/upaas/internal/database"
|
||||
"sneak.berlin/go/upaas/internal/globals"
|
||||
"sneak.berlin/go/upaas/internal/logger"
|
||||
"sneak.berlin/go/upaas/internal/models"
|
||||
"sneak.berlin/go/upaas/internal/service/app"
|
||||
)
|
||||
|
||||
// backupTestContext bundles test dependencies for backup tests.
|
||||
type backupTestContext struct {
|
||||
svc *app.Service
|
||||
db *database.Database
|
||||
}
|
||||
|
||||
func setupBackupTest(t *testing.T) *backupTestContext {
|
||||
t.Helper()
|
||||
|
||||
tmpDir := t.TempDir()
|
||||
|
||||
globals.SetAppname("upaas-test")
|
||||
globals.SetVersion("test")
|
||||
|
||||
globalsInst, err := globals.New(fx.Lifecycle(nil))
|
||||
require.NoError(t, err)
|
||||
|
||||
loggerInst, err := logger.New(
|
||||
fx.Lifecycle(nil),
|
||||
logger.Params{Globals: globalsInst},
|
||||
)
|
||||
require.NoError(t, err)
|
||||
|
||||
cfg := &config.Config{
|
||||
Port: 8080,
|
||||
DataDir: tmpDir,
|
||||
SessionSecret: "test-secret-key-at-least-32-chars",
|
||||
}
|
||||
|
||||
dbInst, err := database.New(fx.Lifecycle(nil), database.Params{
|
||||
Logger: loggerInst,
|
||||
Config: cfg,
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
svc, err := app.New(fx.Lifecycle(nil), app.ServiceParams{
|
||||
Logger: loggerInst,
|
||||
Database: dbInst,
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
return &backupTestContext{svc: svc, db: dbInst}
|
||||
}
|
||||
|
||||
// createAppWithFullConfig creates an app with env vars, labels, volumes, and ports.
|
||||
func createAppWithFullConfig(
|
||||
t *testing.T,
|
||||
btc *backupTestContext,
|
||||
name string,
|
||||
) *models.App {
|
||||
t.Helper()
|
||||
|
||||
createdApp, err := btc.svc.CreateApp(context.Background(), app.CreateAppInput{
|
||||
Name: name,
|
||||
RepoURL: "git@example.com:user/" + name + ".git",
|
||||
Branch: "develop",
|
||||
NtfyTopic: "https://ntfy.sh/" + name,
|
||||
DockerNetwork: "test-network",
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
require.NoError(t, btc.svc.AddEnvVar(
|
||||
context.Background(), createdApp.ID, "DB_HOST", "localhost",
|
||||
))
|
||||
require.NoError(t, btc.svc.AddEnvVar(
|
||||
context.Background(), createdApp.ID, "DB_PORT", "5432",
|
||||
))
|
||||
require.NoError(t, btc.svc.AddLabel(
|
||||
context.Background(), createdApp.ID, "traefik.enable", "true",
|
||||
))
|
||||
require.NoError(t, btc.svc.AddVolume(
|
||||
context.Background(), createdApp.ID, "/data", "/app/data", false,
|
||||
))
|
||||
|
||||
port := models.NewPort(btc.db)
|
||||
port.AppID = createdApp.ID
|
||||
port.HostPort = 9090
|
||||
port.ContainerPort = 8080
|
||||
port.Protocol = models.PortProtocolTCP
|
||||
require.NoError(t, port.Save(context.Background()))
|
||||
|
||||
return createdApp
|
||||
}
|
||||
|
||||
// createAppWithConfigPort creates an app like createAppWithFullConfig but with
|
||||
// a custom host port to avoid UNIQUE constraint collisions.
|
||||
func createAppWithConfigPort(
|
||||
t *testing.T,
|
||||
btc *backupTestContext,
|
||||
name string,
|
||||
hostPort int,
|
||||
) *models.App {
|
||||
t.Helper()
|
||||
|
||||
createdApp, err := btc.svc.CreateApp(context.Background(), app.CreateAppInput{
|
||||
Name: name,
|
||||
RepoURL: "git@example.com:user/" + name + ".git",
|
||||
Branch: "develop",
|
||||
NtfyTopic: "https://ntfy.sh/" + name,
|
||||
DockerNetwork: "test-network",
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
require.NoError(t, btc.svc.AddEnvVar(
|
||||
context.Background(), createdApp.ID, "DB_HOST", "localhost",
|
||||
))
|
||||
require.NoError(t, btc.svc.AddLabel(
|
||||
context.Background(), createdApp.ID, "traefik.enable", "true",
|
||||
))
|
||||
require.NoError(t, btc.svc.AddVolume(
|
||||
context.Background(), createdApp.ID, "/data2", "/app/data2", false,
|
||||
))
|
||||
|
||||
port := models.NewPort(btc.db)
|
||||
port.AppID = createdApp.ID
|
||||
port.HostPort = hostPort
|
||||
port.ContainerPort = 8080
|
||||
port.Protocol = models.PortProtocolTCP
|
||||
require.NoError(t, port.Save(context.Background()))
|
||||
|
||||
return createdApp
|
||||
}
|
||||
|
||||
func TestExportApp(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
btc := setupBackupTest(t)
|
||||
createdApp := createAppWithFullConfig(t, btc, "export-svc-test")
|
||||
|
||||
bundle, err := btc.svc.ExportApp(context.Background(), createdApp)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, 1, bundle.Version)
|
||||
assert.NotEmpty(t, bundle.ExportedAt)
|
||||
require.Len(t, bundle.Apps, 1)
|
||||
|
||||
ab := bundle.Apps[0]
|
||||
assert.Equal(t, "export-svc-test", ab.Name)
|
||||
assert.Equal(t, "develop", ab.Branch)
|
||||
assert.Equal(t, "test-network", ab.DockerNetwork)
|
||||
assert.Equal(t, "https://ntfy.sh/export-svc-test", ab.NtfyTopic)
|
||||
assert.Len(t, ab.EnvVars, 2)
|
||||
assert.Len(t, ab.Labels, 1)
|
||||
assert.Len(t, ab.Volumes, 1)
|
||||
assert.Len(t, ab.Ports, 1)
|
||||
assert.Equal(t, 9090, ab.Ports[0].HostPort)
|
||||
assert.Equal(t, 8080, ab.Ports[0].ContainerPort)
|
||||
assert.Equal(t, "tcp", ab.Ports[0].Protocol)
|
||||
}
|
||||
|
||||
func TestExportAllApps(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
btc := setupBackupTest(t)
|
||||
createAppWithFullConfig(t, btc, "export-all-1")
|
||||
createAppWithConfigPort(t, btc, "export-all-2", 9091)
|
||||
|
||||
bundle, err := btc.svc.ExportAllApps(context.Background())
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, 1, bundle.Version)
|
||||
assert.Len(t, bundle.Apps, 2)
|
||||
}
|
||||
|
||||
func TestExportAllAppsEmpty(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
btc := setupBackupTest(t)
|
||||
|
||||
bundle, err := btc.svc.ExportAllApps(context.Background())
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Empty(t, bundle.Apps)
|
||||
}
|
||||
|
||||
func TestImportApps(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
btc := setupBackupTest(t)
|
||||
|
||||
bundle := &app.BackupBundle{
|
||||
Version: 1,
|
||||
ExportedAt: "2025-01-01T00:00:00Z",
|
||||
Apps: []app.Backup{
|
||||
{
|
||||
Name: "imported-test",
|
||||
RepoURL: "git@example.com:user/imported.git",
|
||||
Branch: "main",
|
||||
DockerfilePath: "Dockerfile",
|
||||
DockerNetwork: "my-network",
|
||||
EnvVars: []app.BackupEnvVar{
|
||||
{Key: "FOO", Value: "bar"},
|
||||
},
|
||||
Labels: []app.BackupLabel{
|
||||
{Key: "app", Value: "test"},
|
||||
},
|
||||
Volumes: []app.BackupVolume{
|
||||
{HostPath: "/host", ContainerPath: "/container", ReadOnly: true},
|
||||
},
|
||||
Ports: []app.BackupPort{
|
||||
{HostPort: 3000, ContainerPort: 8080, Protocol: "tcp"},
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
imported, skipped, err := btc.svc.ImportApps(context.Background(), bundle)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, []string{"imported-test"}, imported)
|
||||
assert.Empty(t, skipped)
|
||||
|
||||
// Verify the app was created
|
||||
apps, _ := btc.svc.ListApps(context.Background())
|
||||
require.Len(t, apps, 1)
|
||||
assert.Equal(t, "imported-test", apps[0].Name)
|
||||
assert.True(t, apps[0].DockerNetwork.Valid)
|
||||
assert.Equal(t, "my-network", apps[0].DockerNetwork.String)
|
||||
|
||||
// Has fresh secrets
|
||||
assert.NotEmpty(t, apps[0].WebhookSecret)
|
||||
assert.NotEmpty(t, apps[0].SSHPublicKey)
|
||||
|
||||
// Verify sub-resources
|
||||
envVars, _ := apps[0].GetEnvVars(context.Background())
|
||||
assert.Len(t, envVars, 1)
|
||||
|
||||
labels, _ := apps[0].GetLabels(context.Background())
|
||||
assert.Len(t, labels, 1)
|
||||
|
||||
volumes, _ := apps[0].GetVolumes(context.Background())
|
||||
assert.Len(t, volumes, 1)
|
||||
assert.True(t, volumes[0].ReadOnly)
|
||||
|
||||
ports, _ := apps[0].GetPorts(context.Background())
|
||||
assert.Len(t, ports, 1)
|
||||
assert.Equal(t, 3000, ports[0].HostPort)
|
||||
}
|
||||
|
||||
func TestImportAppsSkipsDuplicates(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
btc := setupBackupTest(t)
|
||||
|
||||
// Create existing app
|
||||
_, err := btc.svc.CreateApp(context.Background(), app.CreateAppInput{
|
||||
Name: "existing",
|
||||
RepoURL: "git@example.com:user/existing.git",
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
bundle := &app.BackupBundle{
|
||||
Version: 1,
|
||||
ExportedAt: "2025-01-01T00:00:00Z",
|
||||
Apps: []app.Backup{
|
||||
{
|
||||
Name: "existing",
|
||||
RepoURL: "git@example.com:user/existing.git",
|
||||
Branch: "main",
|
||||
DockerfilePath: "Dockerfile",
|
||||
EnvVars: []app.BackupEnvVar{},
|
||||
Labels: []app.BackupLabel{},
|
||||
Volumes: []app.BackupVolume{},
|
||||
Ports: []app.BackupPort{},
|
||||
},
|
||||
{
|
||||
Name: "brand-new",
|
||||
RepoURL: "git@example.com:user/new.git",
|
||||
Branch: "main",
|
||||
DockerfilePath: "Dockerfile",
|
||||
EnvVars: []app.BackupEnvVar{},
|
||||
Labels: []app.BackupLabel{},
|
||||
Volumes: []app.BackupVolume{},
|
||||
Ports: []app.BackupPort{},
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
imported, skipped, err := btc.svc.ImportApps(context.Background(), bundle)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, []string{"brand-new"}, imported)
|
||||
assert.Equal(t, []string{"existing"}, skipped)
|
||||
}
|
||||
|
||||
func TestImportAppsPortDefaultProtocol(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
btc := setupBackupTest(t)
|
||||
|
||||
bundle := &app.BackupBundle{
|
||||
Version: 1,
|
||||
ExportedAt: "2025-01-01T00:00:00Z",
|
||||
Apps: []app.Backup{
|
||||
{
|
||||
Name: "port-default-test",
|
||||
RepoURL: "git@example.com:user/repo.git",
|
||||
Branch: "main",
|
||||
DockerfilePath: "Dockerfile",
|
||||
EnvVars: []app.BackupEnvVar{},
|
||||
Labels: []app.BackupLabel{},
|
||||
Volumes: []app.BackupVolume{},
|
||||
Ports: []app.BackupPort{
|
||||
{HostPort: 80, ContainerPort: 80, Protocol: ""},
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
imported, _, err := btc.svc.ImportApps(context.Background(), bundle)
|
||||
require.NoError(t, err)
|
||||
assert.Len(t, imported, 1)
|
||||
|
||||
apps, _ := btc.svc.ListApps(context.Background())
|
||||
ports, _ := apps[0].GetPorts(context.Background())
|
||||
require.Len(t, ports, 1)
|
||||
assert.Equal(t, models.PortProtocolTCP, ports[0].Protocol)
|
||||
}
|
||||
|
||||
func TestExportImportRoundTripService(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
btc := setupBackupTest(t)
|
||||
createAppWithFullConfig(t, btc, "roundtrip-svc")
|
||||
|
||||
// Export
|
||||
bundle, err := btc.svc.ExportAllApps(context.Background())
|
||||
require.NoError(t, err)
|
||||
require.Len(t, bundle.Apps, 1)
|
||||
|
||||
// Delete original
|
||||
apps, _ := btc.svc.ListApps(context.Background())
|
||||
for _, a := range apps {
|
||||
require.NoError(t, btc.svc.DeleteApp(context.Background(), a))
|
||||
}
|
||||
|
||||
// Import
|
||||
imported, skipped, err := btc.svc.ImportApps(context.Background(), bundle)
|
||||
require.NoError(t, err)
|
||||
assert.Len(t, imported, 1)
|
||||
assert.Empty(t, skipped)
|
||||
|
||||
// Verify round-trip fidelity
|
||||
restored, _ := btc.svc.ListApps(context.Background())
|
||||
require.Len(t, restored, 1)
|
||||
assert.Equal(t, "roundtrip-svc", restored[0].Name)
|
||||
assert.Equal(t, "develop", restored[0].Branch)
|
||||
assert.Equal(t, "test-network", restored[0].DockerNetwork.String)
|
||||
|
||||
envVars, _ := restored[0].GetEnvVars(context.Background())
|
||||
assert.Len(t, envVars, 2)
|
||||
|
||||
labels, _ := restored[0].GetLabels(context.Background())
|
||||
assert.Len(t, labels, 1)
|
||||
|
||||
volumes, _ := restored[0].GetVolumes(context.Background())
|
||||
assert.Len(t, volumes, 1)
|
||||
|
||||
ports, _ := restored[0].GetPorts(context.Background())
|
||||
assert.Len(t, ports, 1)
|
||||
}
|
||||
@@ -1094,14 +1094,28 @@ func (svc *Service) buildContainerOptions(
|
||||
network = app.DockerNetwork.String
|
||||
}
|
||||
|
||||
var cpuLimit float64
|
||||
|
||||
if app.CPULimit.Valid {
|
||||
cpuLimit = app.CPULimit.Float64
|
||||
}
|
||||
|
||||
var memoryLimit int64
|
||||
|
||||
if app.MemoryLimit.Valid {
|
||||
memoryLimit = app.MemoryLimit.Int64
|
||||
}
|
||||
|
||||
return docker.CreateContainerOptions{
|
||||
Name: "upaas-" + app.Name,
|
||||
Image: imageID.String(),
|
||||
Env: envMap,
|
||||
Labels: buildLabelMap(app, labels),
|
||||
Volumes: buildVolumeMounts(volumes),
|
||||
Ports: buildPortMappings(ports),
|
||||
Network: network,
|
||||
Name: "upaas-" + app.Name,
|
||||
Image: imageID.String(),
|
||||
Env: envMap,
|
||||
Labels: buildLabelMap(app, labels),
|
||||
Volumes: buildVolumeMounts(volumes),
|
||||
Ports: buildPortMappings(ports),
|
||||
Network: network,
|
||||
CPULimit: cpuLimit,
|
||||
MemoryLimit: memoryLimit,
|
||||
}, nil
|
||||
}
|
||||
|
||||
|
||||
@@ -2,6 +2,7 @@ package deploy_test
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
"log/slog"
|
||||
"os"
|
||||
"testing"
|
||||
@@ -43,3 +44,93 @@ func TestBuildContainerOptionsUsesImageID(t *testing.T) {
|
||||
t.Errorf("expected Name=%q, got %q", "upaas-myapp", opts.Name)
|
||||
}
|
||||
}
|
||||
|
||||
func TestBuildContainerOptionsNoResourceLimits(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
db := database.NewTestDatabase(t)
|
||||
|
||||
app := models.NewApp(db)
|
||||
app.Name = "nolimits"
|
||||
|
||||
err := app.Save(context.Background())
|
||||
if err != nil {
|
||||
t.Fatalf("failed to save app: %v", err)
|
||||
}
|
||||
|
||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
||||
svc := deploy.NewTestService(log)
|
||||
|
||||
opts, err := svc.BuildContainerOptionsExported(
|
||||
context.Background(), app, docker.ImageID("test:latest"),
|
||||
)
|
||||
if err != nil {
|
||||
t.Fatalf("buildContainerOptions returned error: %v", err)
|
||||
}
|
||||
|
||||
if opts.CPULimit != 0 {
|
||||
t.Errorf("expected CPULimit=0, got %v", opts.CPULimit)
|
||||
}
|
||||
|
||||
if opts.MemoryLimit != 0 {
|
||||
t.Errorf("expected MemoryLimit=0, got %v", opts.MemoryLimit)
|
||||
}
|
||||
}
|
||||
|
||||
func TestBuildContainerOptionsCPULimit(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
db := database.NewTestDatabase(t)
|
||||
|
||||
app := models.NewApp(db)
|
||||
app.Name = "cpulimit"
|
||||
app.CPULimit = sql.NullFloat64{Float64: 0.5, Valid: true}
|
||||
|
||||
err := app.Save(context.Background())
|
||||
if err != nil {
|
||||
t.Fatalf("failed to save app: %v", err)
|
||||
}
|
||||
|
||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
||||
svc := deploy.NewTestService(log)
|
||||
|
||||
opts, err := svc.BuildContainerOptionsExported(
|
||||
context.Background(), app, docker.ImageID("test:latest"),
|
||||
)
|
||||
if err != nil {
|
||||
t.Fatalf("buildContainerOptions returned error: %v", err)
|
||||
}
|
||||
|
||||
if opts.CPULimit != 0.5 {
|
||||
t.Errorf("expected CPULimit=0.5, got %v", opts.CPULimit)
|
||||
}
|
||||
}
|
||||
|
||||
func TestBuildContainerOptionsMemoryLimit(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
db := database.NewTestDatabase(t)
|
||||
|
||||
app := models.NewApp(db)
|
||||
app.Name = "memlimit"
|
||||
app.MemoryLimit = sql.NullInt64{Int64: 536870912, Valid: true} // 512m
|
||||
|
||||
err := app.Save(context.Background())
|
||||
if err != nil {
|
||||
t.Fatalf("failed to save app: %v", err)
|
||||
}
|
||||
|
||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
||||
svc := deploy.NewTestService(log)
|
||||
|
||||
opts, err := svc.BuildContainerOptionsExported(
|
||||
context.Background(), app, docker.ImageID("test:latest"),
|
||||
)
|
||||
if err != nil {
|
||||
t.Fatalf("buildContainerOptions returned error: %v", err)
|
||||
}
|
||||
|
||||
if opts.MemoryLimit != 536870912 {
|
||||
t.Errorf("expected MemoryLimit=536870912, got %v", opts.MemoryLimit)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -432,18 +432,6 @@
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Backup -->
|
||||
<div class="card p-6 mb-6">
|
||||
<h2 class="section-title mb-4">Backup</h2>
|
||||
<p class="text-sm text-gray-500 mb-3">Export this app's configuration (settings, env vars, labels, volumes, ports) as a JSON file for backup or migration.</p>
|
||||
<a href="/apps/{{.App.ID}}/export" class="btn-secondary">
|
||||
<svg class="w-4 h-4 mr-1 inline" fill="none" stroke="currentColor" viewBox="0 0 24 24">
|
||||
<path stroke-linecap="round" stroke-linejoin="round" stroke-width="2" d="M4 16v1a3 3 0 003 3h10a3 3 0 003-3v-1m-4-4l-4 4m0 0l-4-4m4 4V4"/>
|
||||
</svg>
|
||||
Export Config
|
||||
</a>
|
||||
</div>
|
||||
|
||||
<!-- Danger Zone -->
|
||||
<div class="card border-2 border-error-500/20 bg-error-50/50 p-6">
|
||||
<h2 class="text-lg font-medium text-error-700 mb-4">Danger Zone</h2>
|
||||
|
||||
@@ -114,6 +114,38 @@
|
||||
>
|
||||
</div>
|
||||
|
||||
<hr class="border-gray-200">
|
||||
|
||||
<h3 class="text-lg font-medium text-gray-900">Resource Limits</h3>
|
||||
|
||||
<div class="grid grid-cols-2 gap-4">
|
||||
<div class="form-group">
|
||||
<label for="cpu_limit" class="label">CPU Limit (cores)</label>
|
||||
<input
|
||||
type="text"
|
||||
id="cpu_limit"
|
||||
name="cpu_limit"
|
||||
value="{{if .App.CPULimit.Valid}}{{.App.CPULimit.Float64}}{{end}}"
|
||||
class="input"
|
||||
placeholder="e.g. 0.5, 1, 2"
|
||||
>
|
||||
<p class="text-sm text-gray-500 mt-1">Number of CPU cores (e.g. 0.5 = half a core)</p>
|
||||
</div>
|
||||
|
||||
<div class="form-group">
|
||||
<label for="memory_limit" class="label">Memory Limit</label>
|
||||
<input
|
||||
type="text"
|
||||
id="memory_limit"
|
||||
name="memory_limit"
|
||||
value="{{if .App.MemoryLimit.Valid}}{{formatMemoryBytes .App.MemoryLimit.Int64}}{{end}}"
|
||||
class="input"
|
||||
placeholder="e.g. 256m, 1g"
|
||||
>
|
||||
<p class="text-sm text-gray-500 mt-1">Memory with unit suffix (k, m, g) or plain bytes</p>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="flex justify-end gap-3 pt-4">
|
||||
<a href="/apps/{{.App.ID}}" class="btn-secondary">Cancel</a>
|
||||
<button type="submit" class="btn-primary">Save Changes</button>
|
||||
|
||||
@@ -1,62 +0,0 @@
|
||||
{{template "base" .}}
|
||||
|
||||
{{define "title"}}Import Backup - µPaaS{{end}}
|
||||
|
||||
{{define "content"}}
|
||||
{{template "nav" .}}
|
||||
|
||||
<main class="max-w-4xl mx-auto px-4 py-8">
|
||||
<div class="mb-6">
|
||||
<a href="/" class="text-primary-600 hover:text-primary-800 inline-flex items-center">
|
||||
<svg class="w-4 h-4 mr-1" fill="none" stroke="currentColor" viewBox="0 0 24 24">
|
||||
<path stroke-linecap="round" stroke-linejoin="round" stroke-width="2" d="M15 19l-7-7 7-7"/>
|
||||
</svg>
|
||||
Back to Dashboard
|
||||
</a>
|
||||
</div>
|
||||
|
||||
{{template "alert-success" .}}
|
||||
{{template "alert-error" .}}
|
||||
|
||||
<h1 class="text-2xl font-medium text-gray-900 mb-6">Import Backup</h1>
|
||||
|
||||
<div class="card p-6 mb-6">
|
||||
<h2 class="section-title mb-4">Restore from Backup File</h2>
|
||||
<p class="text-sm text-gray-500 mb-4">
|
||||
Upload a previously exported µPaaS backup file (JSON) to restore app configurations.
|
||||
New apps will be created with fresh SSH keys and webhook secrets.
|
||||
Apps whose names already exist will be skipped.
|
||||
</p>
|
||||
<form method="POST" action="/backup/import" enctype="multipart/form-data">
|
||||
{{ .CSRFField }}
|
||||
<div class="mb-4">
|
||||
<label for="backup_file" class="form-label">Backup File</label>
|
||||
<input type="file" id="backup_file" name="backup_file" accept=".json,application/json"
|
||||
class="block w-full text-sm text-gray-500
|
||||
file:mr-4 file:py-2 file:px-4
|
||||
file:rounded file:border-0
|
||||
file:text-sm file:font-medium
|
||||
file:bg-primary-50 file:text-primary-700
|
||||
hover:file:bg-primary-100
|
||||
cursor-pointer">
|
||||
</div>
|
||||
<button type="submit" class="btn-primary">Import</button>
|
||||
</form>
|
||||
</div>
|
||||
|
||||
<div class="card p-6">
|
||||
<h2 class="section-title mb-4">Export All Apps</h2>
|
||||
<p class="text-sm text-gray-500 mb-4">
|
||||
Download a backup of all app configurations. This includes app settings,
|
||||
environment variables, labels, volumes, and port mappings.
|
||||
Secrets (SSH keys, webhook tokens) are not included — they are regenerated on import.
|
||||
</p>
|
||||
<a href="/backup/export" class="btn-secondary">
|
||||
<svg class="w-4 h-4 mr-1 inline" fill="none" stroke="currentColor" viewBox="0 0 24 24">
|
||||
<path stroke-linecap="round" stroke-linejoin="round" stroke-width="2" d="M4 16v1a3 3 0 003 3h10a3 3 0 003-3v-1m-4-4l-4 4m0 0l-4-4m4 4V4"/>
|
||||
</svg>
|
||||
Export All Apps
|
||||
</a>
|
||||
</div>
|
||||
</main>
|
||||
{{end}}
|
||||
@@ -11,20 +11,12 @@
|
||||
|
||||
<div class="section-header">
|
||||
<h1 class="text-2xl font-medium text-gray-900">Applications</h1>
|
||||
<div class="flex gap-3">
|
||||
<a href="/backup/import" class="btn-secondary">
|
||||
<svg class="w-4 h-4 mr-1 inline" fill="none" stroke="currentColor" viewBox="0 0 24 24">
|
||||
<path stroke-linecap="round" stroke-linejoin="round" stroke-width="2" d="M4 16v1a3 3 0 003 3h10a3 3 0 003-3v-1m-4-8l-4-4m0 0L8 8m4-4v12"/>
|
||||
</svg>
|
||||
Backup / Restore
|
||||
</a>
|
||||
<a href="/apps/new" class="btn-primary">
|
||||
<svg class="w-5 h-5 mr-1" fill="none" stroke="currentColor" viewBox="0 0 24 24">
|
||||
<path stroke-linecap="round" stroke-linejoin="round" stroke-width="2" d="M12 4v16m8-8H4"/>
|
||||
</svg>
|
||||
New App
|
||||
</a>
|
||||
</div>
|
||||
<a href="/apps/new" class="btn-primary">
|
||||
<svg class="w-5 h-5 mr-1" fill="none" stroke="currentColor" viewBox="0 0 24 24">
|
||||
<path stroke-linecap="round" stroke-linejoin="round" stroke-width="2" d="M12 4v16m8-8H4"/>
|
||||
</svg>
|
||||
New App
|
||||
</a>
|
||||
</div>
|
||||
|
||||
{{if .AppStats}}
|
||||
|
||||
@@ -6,6 +6,7 @@ import (
|
||||
"fmt"
|
||||
"html/template"
|
||||
"io"
|
||||
"strconv"
|
||||
"sync"
|
||||
)
|
||||
|
||||
@@ -23,6 +24,34 @@ var (
|
||||
templatesMutex sync.RWMutex
|
||||
)
|
||||
|
||||
// templateFuncs returns the custom template function map.
|
||||
func templateFuncs() template.FuncMap {
|
||||
return template.FuncMap{
|
||||
"formatMemoryBytes": formatMemoryBytes,
|
||||
}
|
||||
}
|
||||
|
||||
// Memory unit constants.
|
||||
const (
|
||||
memGigabyte = 1024 * 1024 * 1024
|
||||
memMegabyte = 1024 * 1024
|
||||
memKilobyte = 1024
|
||||
)
|
||||
|
||||
// formatMemoryBytes formats bytes into a human-readable string with unit suffix.
|
||||
func formatMemoryBytes(bytes int64) string {
|
||||
switch {
|
||||
case bytes >= memGigabyte && bytes%memGigabyte == 0:
|
||||
return strconv.FormatInt(bytes/memGigabyte, 10) + "g"
|
||||
case bytes >= memMegabyte && bytes%memMegabyte == 0:
|
||||
return strconv.FormatInt(bytes/memMegabyte, 10) + "m"
|
||||
case bytes >= memKilobyte && bytes%memKilobyte == 0:
|
||||
return strconv.FormatInt(bytes/memKilobyte, 10) + "k"
|
||||
default:
|
||||
return strconv.FormatInt(bytes, 10)
|
||||
}
|
||||
}
|
||||
|
||||
// initTemplates parses base template and creates cloned templates for each page.
|
||||
func initTemplates() {
|
||||
templatesMutex.Lock()
|
||||
@@ -32,8 +61,10 @@ func initTemplates() {
|
||||
return
|
||||
}
|
||||
|
||||
// Parse base template with shared components
|
||||
baseTemplate = template.Must(template.ParseFS(templatesRaw, "base.html"))
|
||||
// Parse base template with shared components and custom functions
|
||||
baseTemplate = template.Must(
|
||||
template.New("base.html").Funcs(templateFuncs()).ParseFS(templatesRaw, "base.html"),
|
||||
)
|
||||
|
||||
// Pages that extend base
|
||||
pages := []string{
|
||||
@@ -45,7 +76,6 @@ func initTemplates() {
|
||||
"app_edit.html",
|
||||
"deployments.html",
|
||||
"webhook_events.html",
|
||||
"backup_import.html",
|
||||
}
|
||||
|
||||
pageTemplates = make(map[string]*template.Template)
|
||||
|
||||
34
templates/templates_test.go
Normal file
34
templates/templates_test.go
Normal file
@@ -0,0 +1,34 @@
|
||||
package templates //nolint:testpackage // tests unexported formatMemoryBytes
|
||||
|
||||
import (
|
||||
"testing"
|
||||
)
|
||||
|
||||
func TestFormatMemoryBytes(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
tests := []struct {
|
||||
name string
|
||||
bytes int64
|
||||
expected string
|
||||
}{
|
||||
{"gigabytes", 1024 * 1024 * 1024, "1g"},
|
||||
{"two gigabytes", 2 * 1024 * 1024 * 1024, "2g"},
|
||||
{"megabytes", 256 * 1024 * 1024, "256m"},
|
||||
{"kilobytes", 512 * 1024, "512k"},
|
||||
{"plain bytes", 12345, "12345"},
|
||||
{"non-even megabytes", 256*1024*1024 + 1, "268435457"},
|
||||
{"zero", 0, "0"},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
got := formatMemoryBytes(tt.bytes)
|
||||
if got != tt.expected {
|
||||
t.Errorf("formatMemoryBytes(%d) = %q, want %q", tt.bytes, got, tt.expected)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user